diff --git a/hbase-handler/pom.xml b/hbase-handler/pom.xml index 45d1f8666d5f..5e85dcc2e0c4 100644 --- a/hbase-handler/pom.xml +++ b/hbase-handler/pom.xml @@ -28,6 +28,10 @@ + + com.google.protobuf + protobuf-java + org.apache.hive @@ -76,9 +80,29 @@ org.apache.hbase - hbase-client + hbase-shaded-client-byo-hadoop provided + + org.apache.hbase + hbase-testing-util + ${hbase.version} + test + + + org.slf4j + slf4j-log4j12 + + + commons-logging + commons-logging + + + org.glassfish.web + javax.servlet.jsp + + + org.apache.hbase hbase-server diff --git a/hbase-handler/src/java/org/apache/hadoop/hive/hbase/HiveHBaseTableSnapshotInputFormat.java b/hbase-handler/src/java/org/apache/hadoop/hive/hbase/HiveHBaseTableSnapshotInputFormat.java index 2d3f34908ab7..afb44d252844 100644 --- a/hbase-handler/src/java/org/apache/hadoop/hive/hbase/HiveHBaseTableSnapshotInputFormat.java +++ b/hbase-handler/src/java/org/apache/hadoop/hive/hbase/HiveHBaseTableSnapshotInputFormat.java @@ -23,8 +23,8 @@ import org.apache.hadoop.hbase.client.Scan; import org.apache.hadoop.hbase.io.ImmutableBytesWritable; import org.apache.hadoop.hbase.mapred.TableSnapshotInputFormat; -import org.apache.hadoop.hbase.protobuf.ProtobufUtil; -import org.apache.hadoop.hbase.protobuf.generated.ClientProtos; +import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; +import org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos; import org.apache.hadoop.mapred.FileInputFormat; import org.apache.hadoop.mapred.InputFormat; import org.apache.hadoop.mapred.InputSplit; diff --git a/hbase-handler/src/java/org/apache/hadoop/hive/hbase/PutWritable.java b/hbase-handler/src/java/org/apache/hadoop/hive/hbase/PutWritable.java index 67ec0e82dbd8..4d108f9ce298 100644 --- a/hbase-handler/src/java/org/apache/hadoop/hive/hbase/PutWritable.java +++ b/hbase-handler/src/java/org/apache/hadoop/hive/hbase/PutWritable.java @@ -27,9 +27,9 @@ import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.KeyValueUtil; import org.apache.hadoop.hbase.client.Put; -import org.apache.hadoop.hbase.protobuf.ProtobufUtil; -import org.apache.hadoop.hbase.protobuf.generated.ClientProtos; -import org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.MutationType; +import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; +import org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos; +import org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.MutationProto.MutationType; import org.apache.hadoop.io.Writable; public class PutWritable implements Writable { diff --git a/hbase-handler/src/java/org/apache/hadoop/hive/hbase/ResultWritable.java b/hbase-handler/src/java/org/apache/hadoop/hive/hbase/ResultWritable.java index 2ab2412bcb30..41ab176a32bd 100644 --- a/hbase-handler/src/java/org/apache/hadoop/hive/hbase/ResultWritable.java +++ b/hbase-handler/src/java/org/apache/hadoop/hive/hbase/ResultWritable.java @@ -26,8 +26,8 @@ import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.KeyValueUtil; import org.apache.hadoop.hbase.client.Result; -import org.apache.hadoop.hbase.protobuf.ProtobufUtil; -import org.apache.hadoop.hbase.protobuf.generated.ClientProtos; +import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; +import org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos; import org.apache.hadoop.io.Writable; public class ResultWritable implements Writable { diff --git a/hbase-handler/src/test/org/apache/hadoop/hive/hbase/TestHBaseQueries.java b/hbase-handler/src/test/org/apache/hadoop/hive/hbase/TestHBaseQueries.java index da69f0887f77..2cddb99e9df9 100644 --- a/hbase-handler/src/test/org/apache/hadoop/hive/hbase/TestHBaseQueries.java +++ b/hbase-handler/src/test/org/apache/hadoop/hive/hbase/TestHBaseQueries.java @@ -25,6 +25,7 @@ import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.HBaseTestingUtility; import org.apache.hadoop.hbase.MiniHBaseCluster; +import org.apache.hadoop.hbase.StartMiniClusterOption; import org.apache.hadoop.hbase.zookeeper.MiniZooKeeperCluster; import org.apache.hadoop.hive.conf.HiveConf; import org.apache.hadoop.hive.ql.DriverFactory; @@ -45,6 +46,8 @@ public class TestHBaseQueries { private final HiveConf baseConf; private IDriver driver; + private static HBaseTestingUtility util; + /** * Test class for running queries using HBase tables. Creates a mini ZK and an HBase test cluster. * Each test method must instantiate a driver object first, using either the baseConf or a modified version of that. @@ -64,9 +67,11 @@ public TestHBaseQueries() throws Exception { // set up HBase baseConf.setBoolean("hbase.netty.nativetransport", false); - HBaseTestingUtility util = new HBaseTestingUtility(baseConf); + util = new HBaseTestingUtility(baseConf); util.setZkCluster(zooKeeperCluster); - miniHBaseCluster = util.startMiniHBaseCluster(1, 1); + miniHBaseCluster = util.startMiniHBaseCluster( StartMiniClusterOption.builder() + .numMasters(1).numRegionServers(1) + .build()); // set up HMS backend DB TestTxnDbUtil.setConfValues(baseConf); @@ -77,7 +82,7 @@ public TestHBaseQueries() throws Exception { @AfterClass public static void tearDown() throws IOException { if (miniHBaseCluster != null) { - miniHBaseCluster.shutdown(); + util.shutdownMiniHBaseCluster(); } if (zooKeeperCluster != null) { zooKeeperCluster.shutdown(); diff --git a/itests/hcatalog-unit/pom.xml b/itests/hcatalog-unit/pom.xml index adb8aa30915b..3f667a111050 100644 --- a/itests/hcatalog-unit/pom.xml +++ b/itests/hcatalog-unit/pom.xml @@ -265,6 +265,12 @@ tests test + + org.apache.hbase + hbase-zookeeper + tests + test + org.apache.pig pig diff --git a/itests/hive-blobstore/pom.xml b/itests/hive-blobstore/pom.xml index 054ebf27a53d..37d16945b1bb 100644 --- a/itests/hive-blobstore/pom.xml +++ b/itests/hive-blobstore/pom.xml @@ -33,7 +33,7 @@ ${hadoop.version} -mkdir -p - 3.3.0 + 3.21.1 diff --git a/itests/hive-minikdc/pom.xml b/itests/hive-minikdc/pom.xml index 856b47f09a25..fdae87df2033 100644 --- a/itests/hive-minikdc/pom.xml +++ b/itests/hive-minikdc/pom.xml @@ -27,7 +27,7 @@ ../.. None - 3.3.0 + 3.21.0 diff --git a/itests/qtest/pom.xml b/itests/qtest/pom.xml index bc58476789df..3167362ad593 100644 --- a/itests/qtest/pom.xml +++ b/itests/qtest/pom.xml @@ -36,6 +36,10 @@ + + com.google.protobuf + protobuf-java + org.apache.hive diff --git a/itests/util/pom.xml b/itests/util/pom.xml index 1215aa20009c..d6a2bcbbf0fd 100644 --- a/itests/util/pom.xml +++ b/itests/util/pom.xml @@ -164,6 +164,11 @@ org.apache.hbase hbase-mapreduce + + org.apache.hbase + hbase-zookeeper + tests + org.apache.kudu kudu-test-utils diff --git a/llap-common/src/gen/protobuf/gen-java/org/apache/hadoop/hive/llap/daemon/rpc/LlapDaemonProtocolProtos.java b/llap-common/src/gen/protobuf/gen-java/org/apache/hadoop/hive/llap/daemon/rpc/LlapDaemonProtocolProtos.java index b50448053bfa..235779bc46d5 100644 --- a/llap-common/src/gen/protobuf/gen-java/org/apache/hadoop/hive/llap/daemon/rpc/LlapDaemonProtocolProtos.java +++ b/llap-common/src/gen/protobuf/gen-java/org/apache/hadoop/hive/llap/daemon/rpc/LlapDaemonProtocolProtos.java @@ -5,8 +5,14 @@ public final class LlapDaemonProtocolProtos { private LlapDaemonProtocolProtos() {} + public static void registerAllExtensions( + com.google.protobuf.ExtensionRegistryLite registry) { + } + public static void registerAllExtensions( com.google.protobuf.ExtensionRegistry registry) { + registerAllExtensions( + (com.google.protobuf.ExtensionRegistryLite) registry); } /** * Protobuf enum {@code SourceStateProto} @@ -16,11 +22,11 @@ public enum SourceStateProto /** * S_SUCCEEDED = 1; */ - S_SUCCEEDED(0, 1), + S_SUCCEEDED(1), /** * S_RUNNING = 2; */ - S_RUNNING(1, 2), + S_RUNNING(2), ; /** @@ -33,9 +39,25 @@ public enum SourceStateProto public static final int S_RUNNING_VALUE = 2; - public final int getNumber() { return value; } + public final int getNumber() { + return value; + } + /** + * @param value The numeric wire value of the corresponding enum entry. + * @return The enum associated with the given numeric wire value. + * @deprecated Use {@link #forNumber(int)} instead. + */ + @java.lang.Deprecated public static SourceStateProto valueOf(int value) { + return forNumber(value); + } + + /** + * @param value The numeric wire value of the corresponding enum entry. + * @return The enum associated with the given numeric wire value. + */ + public static SourceStateProto forNumber(int value) { switch (value) { case 1: return S_SUCCEEDED; case 2: return S_RUNNING; @@ -47,17 +69,17 @@ public static SourceStateProto valueOf(int value) { internalGetValueMap() { return internalValueMap; } - private static com.google.protobuf.Internal.EnumLiteMap - internalValueMap = + private static final com.google.protobuf.Internal.EnumLiteMap< + SourceStateProto> internalValueMap = new com.google.protobuf.Internal.EnumLiteMap() { public SourceStateProto findValueByNumber(int number) { - return SourceStateProto.valueOf(number); + return SourceStateProto.forNumber(number); } }; public final com.google.protobuf.Descriptors.EnumValueDescriptor getValueDescriptor() { - return getDescriptor().getValues().get(index); + return getDescriptor().getValues().get(ordinal()); } public final com.google.protobuf.Descriptors.EnumDescriptor getDescriptorForType() { @@ -79,11 +101,9 @@ public static SourceStateProto valueOf( return VALUES[desc.getIndex()]; } - private final int index; private final int value; - private SourceStateProto(int index, int value) { - this.index = index; + private SourceStateProto(int value) { this.value = value; } @@ -98,15 +118,15 @@ public enum SubmissionStateProto /** * ACCEPTED = 1; */ - ACCEPTED(0, 1), + ACCEPTED(1), /** * REJECTED = 2; */ - REJECTED(1, 2), + REJECTED(2), /** * EVICTED_OTHER = 3; */ - EVICTED_OTHER(2, 3), + EVICTED_OTHER(3), ; /** @@ -123,9 +143,25 @@ public enum SubmissionStateProto public static final int EVICTED_OTHER_VALUE = 3; - public final int getNumber() { return value; } + public final int getNumber() { + return value; + } + /** + * @param value The numeric wire value of the corresponding enum entry. + * @return The enum associated with the given numeric wire value. + * @deprecated Use {@link #forNumber(int)} instead. + */ + @java.lang.Deprecated public static SubmissionStateProto valueOf(int value) { + return forNumber(value); + } + + /** + * @param value The numeric wire value of the corresponding enum entry. + * @return The enum associated with the given numeric wire value. + */ + public static SubmissionStateProto forNumber(int value) { switch (value) { case 1: return ACCEPTED; case 2: return REJECTED; @@ -138,17 +174,17 @@ public static SubmissionStateProto valueOf(int value) { internalGetValueMap() { return internalValueMap; } - private static com.google.protobuf.Internal.EnumLiteMap - internalValueMap = + private static final com.google.protobuf.Internal.EnumLiteMap< + SubmissionStateProto> internalValueMap = new com.google.protobuf.Internal.EnumLiteMap() { public SubmissionStateProto findValueByNumber(int number) { - return SubmissionStateProto.valueOf(number); + return SubmissionStateProto.forNumber(number); } }; public final com.google.protobuf.Descriptors.EnumValueDescriptor getValueDescriptor() { - return getDescriptor().getValues().get(index); + return getDescriptor().getValues().get(ordinal()); } public final com.google.protobuf.Descriptors.EnumDescriptor getDescriptorForType() { @@ -170,37 +206,38 @@ public static SubmissionStateProto valueOf( return VALUES[desc.getIndex()]; } - private final int index; private final int value; - private SubmissionStateProto(int index, int value) { - this.index = index; + private SubmissionStateProto(int value) { this.value = value; } // @@protoc_insertion_point(enum_scope:SubmissionStateProto) } - public interface UserPayloadProtoOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface UserPayloadProtoOrBuilder extends + // @@protoc_insertion_point(interface_extends:UserPayloadProto) + com.google.protobuf.MessageOrBuilder { - // optional bytes user_payload = 1; /** * optional bytes user_payload = 1; + * @return Whether the userPayload field is set. */ boolean hasUserPayload(); /** * optional bytes user_payload = 1; + * @return The userPayload. */ com.google.protobuf.ByteString getUserPayload(); - // optional int32 version = 2; /** * optional int32 version = 2; + * @return Whether the version field is set. */ boolean hasVersion(); /** * optional int32 version = 2; + * @return The version. */ int getVersion(); } @@ -208,35 +245,38 @@ public interface UserPayloadProtoOrBuilder * Protobuf type {@code UserPayloadProto} */ public static final class UserPayloadProto extends - com.google.protobuf.GeneratedMessage - implements UserPayloadProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:UserPayloadProto) + UserPayloadProtoOrBuilder { + private static final long serialVersionUID = 0L; // Use UserPayloadProto.newBuilder() to construct. - private UserPayloadProto(com.google.protobuf.GeneratedMessage.Builder builder) { + private UserPayloadProto(com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private UserPayloadProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final UserPayloadProto defaultInstance; - public static UserPayloadProto getDefaultInstance() { - return defaultInstance; + private UserPayloadProto() { + userPayload_ = com.google.protobuf.ByteString.EMPTY; } - public UserPayloadProto getDefaultInstanceForType() { - return defaultInstance; + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance( + UnusedPrivateParameter unused) { + return new UserPayloadProto(); } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private UserPayloadProto( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); @@ -248,13 +288,6 @@ private UserPayloadProto( case 0: done = true; break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } case 10: { bitField0_ |= 0x00000001; userPayload_ = input.readBytes(); @@ -265,13 +298,20 @@ private UserPayloadProto( version_ = input.readInt32(); break; } + default: { + if (!parseUnknownField( + input, unknownFields, extensionRegistry, tag)) { + done = true; + } + break; + } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); @@ -282,112 +322,95 @@ private UserPayloadProto( return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_UserPayloadProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_UserPayloadProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public UserPayloadProto parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new UserPayloadProto(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - private int bitField0_; - // optional bytes user_payload = 1; public static final int USER_PAYLOAD_FIELD_NUMBER = 1; private com.google.protobuf.ByteString userPayload_; /** * optional bytes user_payload = 1; + * @return Whether the userPayload field is set. */ + @java.lang.Override public boolean hasUserPayload() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional bytes user_payload = 1; + * @return The userPayload. */ + @java.lang.Override public com.google.protobuf.ByteString getUserPayload() { return userPayload_; } - // optional int32 version = 2; public static final int VERSION_FIELD_NUMBER = 2; private int version_; /** * optional int32 version = 2; + * @return Whether the version field is set. */ + @java.lang.Override public boolean hasVersion() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * optional int32 version = 2; + * @return The version. */ + @java.lang.Override public int getVersion() { return version_; } - private void initFields() { - userPayload_ = com.google.protobuf.ByteString.EMPTY; - version_ = 0; - } private byte memoizedIsInitialized = -1; + @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } + @java.lang.Override public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); - if (((bitField0_ & 0x00000001) == 0x00000001)) { + if (((bitField0_ & 0x00000001) != 0)) { output.writeBytes(1, userPayload_); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { + if (((bitField0_ & 0x00000002) != 0)) { output.writeInt32(2, version_); } - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; + @java.lang.Override public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { + if (((bitField0_ & 0x00000001) != 0)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(1, userPayload_); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { + if (((bitField0_ & 0x00000002) != 0)) { size += com.google.protobuf.CodedOutputStream .computeInt32Size(2, version_); } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { @@ -398,30 +421,27 @@ public boolean equals(final java.lang.Object obj) { } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto) obj; - boolean result = true; - result = result && (hasUserPayload() == other.hasUserPayload()); + if (hasUserPayload() != other.hasUserPayload()) return false; if (hasUserPayload()) { - result = result && getUserPayload() - .equals(other.getUserPayload()); + if (!getUserPayload() + .equals(other.getUserPayload())) return false; } - result = result && (hasVersion() == other.hasVersion()); + if (hasVersion() != other.hasVersion()) return false; if (hasVersion()) { - result = result && (getVersion() - == other.getVersion()); + if (getVersion() + != other.getVersion()) return false; } - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; + if (!unknownFields.equals(other.unknownFields)) return false; + return true; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); + hash = (19 * hash) + getDescriptor().hashCode(); if (hasUserPayload()) { hash = (37 * hash) + USER_PAYLOAD_FIELD_NUMBER; hash = (53 * hash) + getUserPayload().hashCode(); @@ -430,11 +450,22 @@ public int hashCode() { hash = (37 * hash) + VERSION_FIELD_NUMBER; hash = (53 * hash) + getVersion(); } - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parseFrom( + java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parseFrom( + java.nio.ByteBuffer data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { @@ -458,46 +489,59 @@ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.Us } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } + @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -505,14 +549,16 @@ protected Builder newBuilderForType( * Protobuf type {@code UserPayloadProto} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder implements + // @@protoc_insertion_point(builder_implements:UserPayloadProto) + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProtoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_UserPayloadProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_UserPayloadProto_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -525,18 +571,16 @@ private Builder() { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { } } - private static Builder create() { - return new Builder(); - } - + @java.lang.Override public Builder clear() { super.clear(); userPayload_ = com.google.protobuf.ByteString.EMPTY; @@ -546,19 +590,18 @@ public Builder clear() { return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - + @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_UserPayloadProto_descriptor; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.getDefaultInstance(); } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto result = buildPartial(); if (!result.isInitialized()) { @@ -567,23 +610,57 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPaylo return result; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { + if (((from_bitField0_ & 0x00000001) != 0)) { to_bitField0_ |= 0x00000001; } result.userPayload_ = userPayload_; - if (((from_bitField0_ & 0x00000002) == 0x00000002)) { + if (((from_bitField0_ & 0x00000002) != 0)) { + result.version_ = version_; to_bitField0_ |= 0x00000002; } - result.version_ = version_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } + @java.lang.Override + public Builder clone() { + return super.clone(); + } + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.setField(field, value); + } + @java.lang.Override + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + @java.lang.Override + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.addRepeatedField(field, value); + } + @java.lang.Override public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto)other); @@ -601,14 +678,17 @@ public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtoc if (other.hasVersion()) { setVersion(other.getVersion()); } - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } + @java.lang.Override public final boolean isInitialized() { return true; } + @java.lang.Override public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) @@ -618,7 +698,7 @@ public Builder mergeFrom( parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -628,22 +708,27 @@ public Builder mergeFrom( } private int bitField0_; - // optional bytes user_payload = 1; private com.google.protobuf.ByteString userPayload_ = com.google.protobuf.ByteString.EMPTY; /** * optional bytes user_payload = 1; + * @return Whether the userPayload field is set. */ + @java.lang.Override public boolean hasUserPayload() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional bytes user_payload = 1; + * @return The userPayload. */ + @java.lang.Override public com.google.protobuf.ByteString getUserPayload() { return userPayload_; } /** * optional bytes user_payload = 1; + * @param value The userPayload to set. + * @return This builder for chaining. */ public Builder setUserPayload(com.google.protobuf.ByteString value) { if (value == null) { @@ -656,6 +741,7 @@ public Builder setUserPayload(com.google.protobuf.ByteString value) { } /** * optional bytes user_payload = 1; + * @return This builder for chaining. */ public Builder clearUserPayload() { bitField0_ = (bitField0_ & ~0x00000001); @@ -664,22 +750,27 @@ public Builder clearUserPayload() { return this; } - // optional int32 version = 2; private int version_ ; /** * optional int32 version = 2; + * @return Whether the version field is set. */ + @java.lang.Override public boolean hasVersion() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * optional int32 version = 2; + * @return The version. */ + @java.lang.Override public int getVersion() { return version_; } /** * optional int32 version = 2; + * @param value The version to set. + * @return This builder for chaining. */ public Builder setVersion(int value) { bitField0_ |= 0x00000002; @@ -689,6 +780,7 @@ public Builder setVersion(int value) { } /** * optional int32 version = 2; + * @return This builder for chaining. */ public Builder clearVersion() { bitField0_ = (bitField0_ & ~0x00000002); @@ -696,43 +788,88 @@ public Builder clearVersion() { onChanged(); return this; } + @java.lang.Override + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:UserPayloadProto) } + // @@protoc_insertion_point(class_scope:UserPayloadProto) + private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto DEFAULT_INSTANCE; static { - defaultInstance = new UserPayloadProto(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto(); + } + + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser + PARSER = new com.google.protobuf.AbstractParser() { + @java.lang.Override + public UserPayloadProto parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new UserPayloadProto(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto getDefaultInstanceForType() { + return DEFAULT_INSTANCE; } - // @@protoc_insertion_point(class_scope:UserPayloadProto) } - public interface EntityDescriptorProtoOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface EntityDescriptorProtoOrBuilder extends + // @@protoc_insertion_point(interface_extends:EntityDescriptorProto) + com.google.protobuf.MessageOrBuilder { - // optional string class_name = 1; /** * optional string class_name = 1; + * @return Whether the className field is set. */ boolean hasClassName(); /** * optional string class_name = 1; + * @return The className. */ java.lang.String getClassName(); /** * optional string class_name = 1; + * @return The bytes for className. */ com.google.protobuf.ByteString getClassNameBytes(); - // optional .UserPayloadProto user_payload = 2; /** * optional .UserPayloadProto user_payload = 2; + * @return Whether the userPayload field is set. */ boolean hasUserPayload(); /** * optional .UserPayloadProto user_payload = 2; + * @return The userPayload. */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto getUserPayload(); /** @@ -740,13 +877,14 @@ public interface EntityDescriptorProtoOrBuilder */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProtoOrBuilder getUserPayloadOrBuilder(); - // optional bytes history_text = 3; /** * optional bytes history_text = 3; + * @return Whether the historyText field is set. */ boolean hasHistoryText(); /** * optional bytes history_text = 3; + * @return The historyText. */ com.google.protobuf.ByteString getHistoryText(); } @@ -754,35 +892,39 @@ public interface EntityDescriptorProtoOrBuilder * Protobuf type {@code EntityDescriptorProto} */ public static final class EntityDescriptorProto extends - com.google.protobuf.GeneratedMessage - implements EntityDescriptorProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:EntityDescriptorProto) + EntityDescriptorProtoOrBuilder { + private static final long serialVersionUID = 0L; // Use EntityDescriptorProto.newBuilder() to construct. - private EntityDescriptorProto(com.google.protobuf.GeneratedMessage.Builder builder) { + private EntityDescriptorProto(com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private EntityDescriptorProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final EntityDescriptorProto defaultInstance; - public static EntityDescriptorProto getDefaultInstance() { - return defaultInstance; + private EntityDescriptorProto() { + className_ = ""; + historyText_ = com.google.protobuf.ByteString.EMPTY; } - public EntityDescriptorProto getDefaultInstanceForType() { - return defaultInstance; + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance( + UnusedPrivateParameter unused) { + return new EntityDescriptorProto(); } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private EntityDescriptorProto( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); @@ -794,21 +936,15 @@ private EntityDescriptorProto( case 0: done = true; break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } case 10: { + com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000001; - className_ = input.readBytes(); + className_ = bs; break; } case 18: { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.Builder subBuilder = null; - if (((bitField0_ & 0x00000002) == 0x00000002)) { + if (((bitField0_ & 0x00000002) != 0)) { subBuilder = userPayload_.toBuilder(); } userPayload_ = input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.PARSER, extensionRegistry); @@ -824,13 +960,20 @@ private EntityDescriptorProto( historyText_ = input.readBytes(); break; } + default: { + if (!parseUnknownField( + input, unknownFields, extensionRegistry, tag)) { + done = true; + } + break; + } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); @@ -841,41 +984,30 @@ private EntityDescriptorProto( return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_EntityDescriptorProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_EntityDescriptorProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public EntityDescriptorProto parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new EntityDescriptorProto(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - private int bitField0_; - // optional string class_name = 1; public static final int CLASS_NAME_FIELD_NUMBER = 1; - private java.lang.Object className_; + private volatile java.lang.Object className_; /** * optional string class_name = 1; + * @return Whether the className field is set. */ + @java.lang.Override public boolean hasClassName() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional string class_name = 1; + * @return The className. */ + @java.lang.Override public java.lang.String getClassName() { java.lang.Object ref = className_; if (ref instanceof java.lang.String) { @@ -892,7 +1024,9 @@ public java.lang.String getClassName() { } /** * optional string class_name = 1; + * @return The bytes for className. */ + @java.lang.Override public com.google.protobuf.ByteString getClassNameBytes() { java.lang.Object ref = className_; @@ -907,103 +1041,99 @@ public java.lang.String getClassName() { } } - // optional .UserPayloadProto user_payload = 2; public static final int USER_PAYLOAD_FIELD_NUMBER = 2; private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto userPayload_; /** * optional .UserPayloadProto user_payload = 2; + * @return Whether the userPayload field is set. */ + @java.lang.Override public boolean hasUserPayload() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * optional .UserPayloadProto user_payload = 2; + * @return The userPayload. */ + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto getUserPayload() { - return userPayload_; + return userPayload_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.getDefaultInstance() : userPayload_; } /** * optional .UserPayloadProto user_payload = 2; */ + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProtoOrBuilder getUserPayloadOrBuilder() { - return userPayload_; + return userPayload_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.getDefaultInstance() : userPayload_; } - // optional bytes history_text = 3; public static final int HISTORY_TEXT_FIELD_NUMBER = 3; private com.google.protobuf.ByteString historyText_; /** * optional bytes history_text = 3; + * @return Whether the historyText field is set. */ + @java.lang.Override public boolean hasHistoryText() { - return ((bitField0_ & 0x00000004) == 0x00000004); + return ((bitField0_ & 0x00000004) != 0); } /** * optional bytes history_text = 3; + * @return The historyText. */ + @java.lang.Override public com.google.protobuf.ByteString getHistoryText() { return historyText_; } - private void initFields() { - className_ = ""; - userPayload_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.getDefaultInstance(); - historyText_ = com.google.protobuf.ByteString.EMPTY; - } private byte memoizedIsInitialized = -1; + @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } + @java.lang.Override public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); - if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeBytes(1, getClassNameBytes()); + if (((bitField0_ & 0x00000001) != 0)) { + com.google.protobuf.GeneratedMessageV3.writeString(output, 1, className_); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - output.writeMessage(2, userPayload_); + if (((bitField0_ & 0x00000002) != 0)) { + output.writeMessage(2, getUserPayload()); } - if (((bitField0_ & 0x00000004) == 0x00000004)) { + if (((bitField0_ & 0x00000004) != 0)) { output.writeBytes(3, historyText_); } - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; + @java.lang.Override public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(1, getClassNameBytes()); + if (((bitField0_ & 0x00000001) != 0)) { + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, className_); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { + if (((bitField0_ & 0x00000002) != 0)) { size += com.google.protobuf.CodedOutputStream - .computeMessageSize(2, userPayload_); + .computeMessageSize(2, getUserPayload()); } - if (((bitField0_ & 0x00000004) == 0x00000004)) { + if (((bitField0_ & 0x00000004) != 0)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(3, historyText_); } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { @@ -1014,35 +1144,32 @@ public boolean equals(final java.lang.Object obj) { } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto) obj; - boolean result = true; - result = result && (hasClassName() == other.hasClassName()); + if (hasClassName() != other.hasClassName()) return false; if (hasClassName()) { - result = result && getClassName() - .equals(other.getClassName()); + if (!getClassName() + .equals(other.getClassName())) return false; } - result = result && (hasUserPayload() == other.hasUserPayload()); + if (hasUserPayload() != other.hasUserPayload()) return false; if (hasUserPayload()) { - result = result && getUserPayload() - .equals(other.getUserPayload()); + if (!getUserPayload() + .equals(other.getUserPayload())) return false; } - result = result && (hasHistoryText() == other.hasHistoryText()); + if (hasHistoryText() != other.hasHistoryText()) return false; if (hasHistoryText()) { - result = result && getHistoryText() - .equals(other.getHistoryText()); + if (!getHistoryText() + .equals(other.getHistoryText())) return false; } - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; + if (!unknownFields.equals(other.unknownFields)) return false; + return true; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); + hash = (19 * hash) + getDescriptor().hashCode(); if (hasClassName()) { hash = (37 * hash) + CLASS_NAME_FIELD_NUMBER; hash = (53 * hash) + getClassName().hashCode(); @@ -1055,11 +1182,22 @@ public int hashCode() { hash = (37 * hash) + HISTORY_TEXT_FIELD_NUMBER; hash = (53 * hash) + getHistoryText().hashCode(); } - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parseFrom( + java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parseFrom( + java.nio.ByteBuffer data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { @@ -1083,46 +1221,59 @@ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.En } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } + @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -1130,14 +1281,16 @@ protected Builder newBuilderForType( * Protobuf type {@code EntityDescriptorProto} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder implements + // @@protoc_insertion_point(builder_implements:EntityDescriptorProto) + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_EntityDescriptorProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_EntityDescriptorProto_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -1150,25 +1303,23 @@ private Builder() { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { getUserPayloadFieldBuilder(); } } - private static Builder create() { - return new Builder(); - } - + @java.lang.Override public Builder clear() { super.clear(); className_ = ""; bitField0_ = (bitField0_ & ~0x00000001); if (userPayloadBuilder_ == null) { - userPayload_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.getDefaultInstance(); + userPayload_ = null; } else { userPayloadBuilder_.clear(); } @@ -1178,19 +1329,18 @@ public Builder clear() { return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - + @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_EntityDescriptorProto_descriptor; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance(); } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto result = buildPartial(); if (!result.isInitialized()) { @@ -1199,23 +1349,24 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDes return result; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { + if (((from_bitField0_ & 0x00000001) != 0)) { to_bitField0_ |= 0x00000001; } result.className_ = className_; - if (((from_bitField0_ & 0x00000002) == 0x00000002)) { + if (((from_bitField0_ & 0x00000002) != 0)) { + if (userPayloadBuilder_ == null) { + result.userPayload_ = userPayload_; + } else { + result.userPayload_ = userPayloadBuilder_.build(); + } to_bitField0_ |= 0x00000002; } - if (userPayloadBuilder_ == null) { - result.userPayload_ = userPayload_; - } else { - result.userPayload_ = userPayloadBuilder_.build(); - } - if (((from_bitField0_ & 0x00000004) == 0x00000004)) { + if (((from_bitField0_ & 0x00000004) != 0)) { to_bitField0_ |= 0x00000004; } result.historyText_ = historyText_; @@ -1224,6 +1375,39 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDes return result; } + @java.lang.Override + public Builder clone() { + return super.clone(); + } + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.setField(field, value); + } + @java.lang.Override + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + @java.lang.Override + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.addRepeatedField(field, value); + } + @java.lang.Override public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto)other); @@ -1246,14 +1430,17 @@ public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtoc if (other.hasHistoryText()) { setHistoryText(other.getHistoryText()); } - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } + @java.lang.Override public final boolean isInitialized() { return true; } + @java.lang.Override public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) @@ -1263,7 +1450,7 @@ public Builder mergeFrom( parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -1273,23 +1460,27 @@ public Builder mergeFrom( } private int bitField0_; - // optional string class_name = 1; private java.lang.Object className_ = ""; /** * optional string class_name = 1; + * @return Whether the className field is set. */ public boolean hasClassName() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional string class_name = 1; + * @return The className. */ public java.lang.String getClassName() { java.lang.Object ref = className_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - className_ = s; + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + className_ = s; + } return s; } else { return (java.lang.String) ref; @@ -1297,6 +1488,7 @@ public java.lang.String getClassName() { } /** * optional string class_name = 1; + * @return The bytes for className. */ public com.google.protobuf.ByteString getClassNameBytes() { @@ -1313,6 +1505,8 @@ public java.lang.String getClassName() { } /** * optional string class_name = 1; + * @param value The className to set. + * @return This builder for chaining. */ public Builder setClassName( java.lang.String value) { @@ -1326,6 +1520,7 @@ public Builder setClassName( } /** * optional string class_name = 1; + * @return This builder for chaining. */ public Builder clearClassName() { bitField0_ = (bitField0_ & ~0x00000001); @@ -1335,6 +1530,8 @@ public Builder clearClassName() { } /** * optional string class_name = 1; + * @param value The bytes for className to set. + * @return This builder for chaining. */ public Builder setClassNameBytes( com.google.protobuf.ByteString value) { @@ -1347,22 +1544,23 @@ public Builder setClassNameBytes( return this; } - // optional .UserPayloadProto user_payload = 2; - private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto userPayload_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.getDefaultInstance(); - private com.google.protobuf.SingleFieldBuilder< + private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto userPayload_; + private com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProtoOrBuilder> userPayloadBuilder_; /** * optional .UserPayloadProto user_payload = 2; + * @return Whether the userPayload field is set. */ public boolean hasUserPayload() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * optional .UserPayloadProto user_payload = 2; + * @return The userPayload. */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto getUserPayload() { if (userPayloadBuilder_ == null) { - return userPayload_; + return userPayload_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.getDefaultInstance() : userPayload_; } else { return userPayloadBuilder_.getMessage(); } @@ -1402,7 +1600,8 @@ public Builder setUserPayload( */ public Builder mergeUserPayload(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto value) { if (userPayloadBuilder_ == null) { - if (((bitField0_ & 0x00000002) == 0x00000002) && + if (((bitField0_ & 0x00000002) != 0) && + userPayload_ != null && userPayload_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.getDefaultInstance()) { userPayload_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.newBuilder(userPayload_).mergeFrom(value).buildPartial(); @@ -1421,7 +1620,7 @@ public Builder mergeUserPayload(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemo */ public Builder clearUserPayload() { if (userPayloadBuilder_ == null) { - userPayload_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.getDefaultInstance(); + userPayload_ = null; onChanged(); } else { userPayloadBuilder_.clear(); @@ -1444,19 +1643,20 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPaylo if (userPayloadBuilder_ != null) { return userPayloadBuilder_.getMessageOrBuilder(); } else { - return userPayload_; + return userPayload_ == null ? + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.getDefaultInstance() : userPayload_; } } /** * optional .UserPayloadProto user_payload = 2; */ - private com.google.protobuf.SingleFieldBuilder< + private com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProtoOrBuilder> getUserPayloadFieldBuilder() { if (userPayloadBuilder_ == null) { - userPayloadBuilder_ = new com.google.protobuf.SingleFieldBuilder< + userPayloadBuilder_ = new com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPayloadProtoOrBuilder>( - userPayload_, + getUserPayload(), getParentForChildren(), isClean()); userPayload_ = null; @@ -1464,22 +1664,27 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UserPaylo return userPayloadBuilder_; } - // optional bytes history_text = 3; private com.google.protobuf.ByteString historyText_ = com.google.protobuf.ByteString.EMPTY; /** * optional bytes history_text = 3; + * @return Whether the historyText field is set. */ + @java.lang.Override public boolean hasHistoryText() { - return ((bitField0_ & 0x00000004) == 0x00000004); + return ((bitField0_ & 0x00000004) != 0); } /** * optional bytes history_text = 3; + * @return The historyText. */ + @java.lang.Override public com.google.protobuf.ByteString getHistoryText() { return historyText_; } /** * optional bytes history_text = 3; + * @param value The historyText to set. + * @return This builder for chaining. */ public Builder setHistoryText(com.google.protobuf.ByteString value) { if (value == null) { @@ -1492,6 +1697,7 @@ public Builder setHistoryText(com.google.protobuf.ByteString value) { } /** * optional bytes history_text = 3; + * @return This builder for chaining. */ public Builder clearHistoryText() { bitField0_ = (bitField0_ & ~0x00000004); @@ -1499,43 +1705,88 @@ public Builder clearHistoryText() { onChanged(); return this; } + @java.lang.Override + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:EntityDescriptorProto) } + // @@protoc_insertion_point(class_scope:EntityDescriptorProto) + private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto DEFAULT_INSTANCE; static { - defaultInstance = new EntityDescriptorProto(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto(); + } + + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser + PARSER = new com.google.protobuf.AbstractParser() { + @java.lang.Override + public EntityDescriptorProto parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new EntityDescriptorProto(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getDefaultInstanceForType() { + return DEFAULT_INSTANCE; } - // @@protoc_insertion_point(class_scope:EntityDescriptorProto) } - public interface IOSpecProtoOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface IOSpecProtoOrBuilder extends + // @@protoc_insertion_point(interface_extends:IOSpecProto) + com.google.protobuf.MessageOrBuilder { - // optional string connected_vertex_name = 1; /** * optional string connected_vertex_name = 1; + * @return Whether the connectedVertexName field is set. */ boolean hasConnectedVertexName(); /** * optional string connected_vertex_name = 1; + * @return The connectedVertexName. */ java.lang.String getConnectedVertexName(); /** * optional string connected_vertex_name = 1; + * @return The bytes for connectedVertexName. */ com.google.protobuf.ByteString getConnectedVertexNameBytes(); - // optional .EntityDescriptorProto io_descriptor = 2; /** * optional .EntityDescriptorProto io_descriptor = 2; + * @return Whether the ioDescriptor field is set. */ boolean hasIoDescriptor(); /** * optional .EntityDescriptorProto io_descriptor = 2; + * @return The ioDescriptor. */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getIoDescriptor(); /** @@ -1543,13 +1794,14 @@ public interface IOSpecProtoOrBuilder */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder getIoDescriptorOrBuilder(); - // optional int32 physical_edge_count = 3; /** * optional int32 physical_edge_count = 3; + * @return Whether the physicalEdgeCount field is set. */ boolean hasPhysicalEdgeCount(); /** * optional int32 physical_edge_count = 3; + * @return The physicalEdgeCount. */ int getPhysicalEdgeCount(); } @@ -1557,35 +1809,38 @@ public interface IOSpecProtoOrBuilder * Protobuf type {@code IOSpecProto} */ public static final class IOSpecProto extends - com.google.protobuf.GeneratedMessage - implements IOSpecProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:IOSpecProto) + IOSpecProtoOrBuilder { + private static final long serialVersionUID = 0L; // Use IOSpecProto.newBuilder() to construct. - private IOSpecProto(com.google.protobuf.GeneratedMessage.Builder builder) { + private IOSpecProto(com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private IOSpecProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final IOSpecProto defaultInstance; - public static IOSpecProto getDefaultInstance() { - return defaultInstance; + private IOSpecProto() { + connectedVertexName_ = ""; } - public IOSpecProto getDefaultInstanceForType() { - return defaultInstance; + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance( + UnusedPrivateParameter unused) { + return new IOSpecProto(); } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private IOSpecProto( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); @@ -1597,21 +1852,15 @@ private IOSpecProto( case 0: done = true; break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } case 10: { + com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000001; - connectedVertexName_ = input.readBytes(); + connectedVertexName_ = bs; break; } case 18: { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder subBuilder = null; - if (((bitField0_ & 0x00000002) == 0x00000002)) { + if (((bitField0_ & 0x00000002) != 0)) { subBuilder = ioDescriptor_.toBuilder(); } ioDescriptor_ = input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.PARSER, extensionRegistry); @@ -1627,13 +1876,20 @@ private IOSpecProto( physicalEdgeCount_ = input.readInt32(); break; } + default: { + if (!parseUnknownField( + input, unknownFields, extensionRegistry, tag)) { + done = true; + } + break; + } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); @@ -1644,41 +1900,30 @@ private IOSpecProto( return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_IOSpecProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_IOSpecProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public IOSpecProto parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new IOSpecProto(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - private int bitField0_; - // optional string connected_vertex_name = 1; public static final int CONNECTED_VERTEX_NAME_FIELD_NUMBER = 1; - private java.lang.Object connectedVertexName_; + private volatile java.lang.Object connectedVertexName_; /** * optional string connected_vertex_name = 1; + * @return Whether the connectedVertexName field is set. */ + @java.lang.Override public boolean hasConnectedVertexName() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional string connected_vertex_name = 1; + * @return The connectedVertexName. */ + @java.lang.Override public java.lang.String getConnectedVertexName() { java.lang.Object ref = connectedVertexName_; if (ref instanceof java.lang.String) { @@ -1695,7 +1940,9 @@ public java.lang.String getConnectedVertexName() { } /** * optional string connected_vertex_name = 1; + * @return The bytes for connectedVertexName. */ + @java.lang.Override public com.google.protobuf.ByteString getConnectedVertexNameBytes() { java.lang.Object ref = connectedVertexName_; @@ -1710,103 +1957,99 @@ public java.lang.String getConnectedVertexName() { } } - // optional .EntityDescriptorProto io_descriptor = 2; public static final int IO_DESCRIPTOR_FIELD_NUMBER = 2; private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto ioDescriptor_; /** * optional .EntityDescriptorProto io_descriptor = 2; + * @return Whether the ioDescriptor field is set. */ + @java.lang.Override public boolean hasIoDescriptor() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * optional .EntityDescriptorProto io_descriptor = 2; + * @return The ioDescriptor. */ + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getIoDescriptor() { - return ioDescriptor_; + return ioDescriptor_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance() : ioDescriptor_; } /** * optional .EntityDescriptorProto io_descriptor = 2; */ + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder getIoDescriptorOrBuilder() { - return ioDescriptor_; + return ioDescriptor_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance() : ioDescriptor_; } - // optional int32 physical_edge_count = 3; public static final int PHYSICAL_EDGE_COUNT_FIELD_NUMBER = 3; private int physicalEdgeCount_; /** * optional int32 physical_edge_count = 3; + * @return Whether the physicalEdgeCount field is set. */ + @java.lang.Override public boolean hasPhysicalEdgeCount() { - return ((bitField0_ & 0x00000004) == 0x00000004); + return ((bitField0_ & 0x00000004) != 0); } /** * optional int32 physical_edge_count = 3; + * @return The physicalEdgeCount. */ + @java.lang.Override public int getPhysicalEdgeCount() { return physicalEdgeCount_; } - private void initFields() { - connectedVertexName_ = ""; - ioDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance(); - physicalEdgeCount_ = 0; - } private byte memoizedIsInitialized = -1; + @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } + @java.lang.Override public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); - if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeBytes(1, getConnectedVertexNameBytes()); + if (((bitField0_ & 0x00000001) != 0)) { + com.google.protobuf.GeneratedMessageV3.writeString(output, 1, connectedVertexName_); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - output.writeMessage(2, ioDescriptor_); + if (((bitField0_ & 0x00000002) != 0)) { + output.writeMessage(2, getIoDescriptor()); } - if (((bitField0_ & 0x00000004) == 0x00000004)) { + if (((bitField0_ & 0x00000004) != 0)) { output.writeInt32(3, physicalEdgeCount_); } - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; + @java.lang.Override public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(1, getConnectedVertexNameBytes()); + if (((bitField0_ & 0x00000001) != 0)) { + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, connectedVertexName_); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { + if (((bitField0_ & 0x00000002) != 0)) { size += com.google.protobuf.CodedOutputStream - .computeMessageSize(2, ioDescriptor_); + .computeMessageSize(2, getIoDescriptor()); } - if (((bitField0_ & 0x00000004) == 0x00000004)) { + if (((bitField0_ & 0x00000004) != 0)) { size += com.google.protobuf.CodedOutputStream .computeInt32Size(3, physicalEdgeCount_); } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { @@ -1817,35 +2060,32 @@ public boolean equals(final java.lang.Object obj) { } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto) obj; - boolean result = true; - result = result && (hasConnectedVertexName() == other.hasConnectedVertexName()); + if (hasConnectedVertexName() != other.hasConnectedVertexName()) return false; if (hasConnectedVertexName()) { - result = result && getConnectedVertexName() - .equals(other.getConnectedVertexName()); + if (!getConnectedVertexName() + .equals(other.getConnectedVertexName())) return false; } - result = result && (hasIoDescriptor() == other.hasIoDescriptor()); + if (hasIoDescriptor() != other.hasIoDescriptor()) return false; if (hasIoDescriptor()) { - result = result && getIoDescriptor() - .equals(other.getIoDescriptor()); + if (!getIoDescriptor() + .equals(other.getIoDescriptor())) return false; } - result = result && (hasPhysicalEdgeCount() == other.hasPhysicalEdgeCount()); + if (hasPhysicalEdgeCount() != other.hasPhysicalEdgeCount()) return false; if (hasPhysicalEdgeCount()) { - result = result && (getPhysicalEdgeCount() - == other.getPhysicalEdgeCount()); + if (getPhysicalEdgeCount() + != other.getPhysicalEdgeCount()) return false; } - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; + if (!unknownFields.equals(other.unknownFields)) return false; + return true; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); + hash = (19 * hash) + getDescriptor().hashCode(); if (hasConnectedVertexName()) { hash = (37 * hash) + CONNECTED_VERTEX_NAME_FIELD_NUMBER; hash = (53 * hash) + getConnectedVertexName().hashCode(); @@ -1858,11 +2098,22 @@ public int hashCode() { hash = (37 * hash) + PHYSICAL_EDGE_COUNT_FIELD_NUMBER; hash = (53 * hash) + getPhysicalEdgeCount(); } - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parseFrom( + java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parseFrom( + java.nio.ByteBuffer data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { @@ -1886,46 +2137,59 @@ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IO } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } + @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -1933,14 +2197,16 @@ protected Builder newBuilderForType( * Protobuf type {@code IOSpecProto} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder implements + // @@protoc_insertion_point(builder_implements:IOSpecProto) + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_IOSpecProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_IOSpecProto_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -1953,25 +2219,23 @@ private Builder() { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { getIoDescriptorFieldBuilder(); } } - private static Builder create() { - return new Builder(); - } - + @java.lang.Override public Builder clear() { super.clear(); connectedVertexName_ = ""; bitField0_ = (bitField0_ & ~0x00000001); if (ioDescriptorBuilder_ == null) { - ioDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance(); + ioDescriptor_ = null; } else { ioDescriptorBuilder_.clear(); } @@ -1981,19 +2245,18 @@ public Builder clear() { return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - + @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_IOSpecProto_descriptor; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.getDefaultInstance(); } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto result = buildPartial(); if (!result.isInitialized()) { @@ -2002,31 +2265,65 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecPro return result; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { + if (((from_bitField0_ & 0x00000001) != 0)) { to_bitField0_ |= 0x00000001; } result.connectedVertexName_ = connectedVertexName_; - if (((from_bitField0_ & 0x00000002) == 0x00000002)) { + if (((from_bitField0_ & 0x00000002) != 0)) { + if (ioDescriptorBuilder_ == null) { + result.ioDescriptor_ = ioDescriptor_; + } else { + result.ioDescriptor_ = ioDescriptorBuilder_.build(); + } to_bitField0_ |= 0x00000002; } - if (ioDescriptorBuilder_ == null) { - result.ioDescriptor_ = ioDescriptor_; - } else { - result.ioDescriptor_ = ioDescriptorBuilder_.build(); - } - if (((from_bitField0_ & 0x00000004) == 0x00000004)) { + if (((from_bitField0_ & 0x00000004) != 0)) { + result.physicalEdgeCount_ = physicalEdgeCount_; to_bitField0_ |= 0x00000004; } - result.physicalEdgeCount_ = physicalEdgeCount_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } + @java.lang.Override + public Builder clone() { + return super.clone(); + } + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.setField(field, value); + } + @java.lang.Override + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + @java.lang.Override + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.addRepeatedField(field, value); + } + @java.lang.Override public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto)other); @@ -2049,14 +2346,17 @@ public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtoc if (other.hasPhysicalEdgeCount()) { setPhysicalEdgeCount(other.getPhysicalEdgeCount()); } - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } + @java.lang.Override public final boolean isInitialized() { return true; } + @java.lang.Override public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) @@ -2066,7 +2366,7 @@ public Builder mergeFrom( parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -2076,23 +2376,27 @@ public Builder mergeFrom( } private int bitField0_; - // optional string connected_vertex_name = 1; private java.lang.Object connectedVertexName_ = ""; /** * optional string connected_vertex_name = 1; + * @return Whether the connectedVertexName field is set. */ public boolean hasConnectedVertexName() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional string connected_vertex_name = 1; + * @return The connectedVertexName. */ public java.lang.String getConnectedVertexName() { java.lang.Object ref = connectedVertexName_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - connectedVertexName_ = s; + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + connectedVertexName_ = s; + } return s; } else { return (java.lang.String) ref; @@ -2100,6 +2404,7 @@ public java.lang.String getConnectedVertexName() { } /** * optional string connected_vertex_name = 1; + * @return The bytes for connectedVertexName. */ public com.google.protobuf.ByteString getConnectedVertexNameBytes() { @@ -2116,6 +2421,8 @@ public java.lang.String getConnectedVertexName() { } /** * optional string connected_vertex_name = 1; + * @param value The connectedVertexName to set. + * @return This builder for chaining. */ public Builder setConnectedVertexName( java.lang.String value) { @@ -2129,6 +2436,7 @@ public Builder setConnectedVertexName( } /** * optional string connected_vertex_name = 1; + * @return This builder for chaining. */ public Builder clearConnectedVertexName() { bitField0_ = (bitField0_ & ~0x00000001); @@ -2138,6 +2446,8 @@ public Builder clearConnectedVertexName() { } /** * optional string connected_vertex_name = 1; + * @param value The bytes for connectedVertexName to set. + * @return This builder for chaining. */ public Builder setConnectedVertexNameBytes( com.google.protobuf.ByteString value) { @@ -2150,22 +2460,23 @@ public Builder setConnectedVertexNameBytes( return this; } - // optional .EntityDescriptorProto io_descriptor = 2; - private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto ioDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance(); - private com.google.protobuf.SingleFieldBuilder< + private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto ioDescriptor_; + private com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder> ioDescriptorBuilder_; /** * optional .EntityDescriptorProto io_descriptor = 2; + * @return Whether the ioDescriptor field is set. */ public boolean hasIoDescriptor() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * optional .EntityDescriptorProto io_descriptor = 2; + * @return The ioDescriptor. */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getIoDescriptor() { if (ioDescriptorBuilder_ == null) { - return ioDescriptor_; + return ioDescriptor_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance() : ioDescriptor_; } else { return ioDescriptorBuilder_.getMessage(); } @@ -2205,7 +2516,8 @@ public Builder setIoDescriptor( */ public Builder mergeIoDescriptor(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto value) { if (ioDescriptorBuilder_ == null) { - if (((bitField0_ & 0x00000002) == 0x00000002) && + if (((bitField0_ & 0x00000002) != 0) && + ioDescriptor_ != null && ioDescriptor_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance()) { ioDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.newBuilder(ioDescriptor_).mergeFrom(value).buildPartial(); @@ -2224,7 +2536,7 @@ public Builder mergeIoDescriptor(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaem */ public Builder clearIoDescriptor() { if (ioDescriptorBuilder_ == null) { - ioDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance(); + ioDescriptor_ = null; onChanged(); } else { ioDescriptorBuilder_.clear(); @@ -2247,19 +2559,20 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDes if (ioDescriptorBuilder_ != null) { return ioDescriptorBuilder_.getMessageOrBuilder(); } else { - return ioDescriptor_; + return ioDescriptor_ == null ? + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance() : ioDescriptor_; } } /** * optional .EntityDescriptorProto io_descriptor = 2; */ - private com.google.protobuf.SingleFieldBuilder< + private com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder> getIoDescriptorFieldBuilder() { if (ioDescriptorBuilder_ == null) { - ioDescriptorBuilder_ = new com.google.protobuf.SingleFieldBuilder< + ioDescriptorBuilder_ = new com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder>( - ioDescriptor_, + getIoDescriptor(), getParentForChildren(), isClean()); ioDescriptor_ = null; @@ -2267,22 +2580,27 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDes return ioDescriptorBuilder_; } - // optional int32 physical_edge_count = 3; private int physicalEdgeCount_ ; /** * optional int32 physical_edge_count = 3; + * @return Whether the physicalEdgeCount field is set. */ + @java.lang.Override public boolean hasPhysicalEdgeCount() { - return ((bitField0_ & 0x00000004) == 0x00000004); + return ((bitField0_ & 0x00000004) != 0); } /** * optional int32 physical_edge_count = 3; + * @return The physicalEdgeCount. */ + @java.lang.Override public int getPhysicalEdgeCount() { return physicalEdgeCount_; } /** * optional int32 physical_edge_count = 3; + * @param value The physicalEdgeCount to set. + * @return This builder for chaining. */ public Builder setPhysicalEdgeCount(int value) { bitField0_ |= 0x00000004; @@ -2292,6 +2610,7 @@ public Builder setPhysicalEdgeCount(int value) { } /** * optional int32 physical_edge_count = 3; + * @return This builder for chaining. */ public Builder clearPhysicalEdgeCount() { bitField0_ = (bitField0_ & ~0x00000004); @@ -2299,63 +2618,113 @@ public Builder clearPhysicalEdgeCount() { onChanged(); return this; } + @java.lang.Override + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:IOSpecProto) } + // @@protoc_insertion_point(class_scope:IOSpecProto) + private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto DEFAULT_INSTANCE; static { - defaultInstance = new IOSpecProto(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto(); + } + + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser + PARSER = new com.google.protobuf.AbstractParser() { + @java.lang.Override + public IOSpecProto parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new IOSpecProto(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto getDefaultInstanceForType() { + return DEFAULT_INSTANCE; } - // @@protoc_insertion_point(class_scope:IOSpecProto) } - public interface GroupInputSpecProtoOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface GroupInputSpecProtoOrBuilder extends + // @@protoc_insertion_point(interface_extends:GroupInputSpecProto) + com.google.protobuf.MessageOrBuilder { - // optional string group_name = 1; /** * optional string group_name = 1; + * @return Whether the groupName field is set. */ boolean hasGroupName(); /** * optional string group_name = 1; + * @return The groupName. */ java.lang.String getGroupName(); /** * optional string group_name = 1; + * @return The bytes for groupName. */ com.google.protobuf.ByteString getGroupNameBytes(); - // repeated string group_vertices = 2; /** * repeated string group_vertices = 2; + * @return A list containing the groupVertices. */ java.util.List - getGroupVerticesList(); + getGroupVerticesList(); /** * repeated string group_vertices = 2; + * @return The count of groupVertices. */ int getGroupVerticesCount(); /** * repeated string group_vertices = 2; + * @param index The index of the element to return. + * @return The groupVertices at the given index. */ java.lang.String getGroupVertices(int index); /** * repeated string group_vertices = 2; + * @param index The index of the value to return. + * @return The bytes of the groupVertices at the given index. */ com.google.protobuf.ByteString getGroupVerticesBytes(int index); - // optional .EntityDescriptorProto merged_input_descriptor = 3; /** * optional .EntityDescriptorProto merged_input_descriptor = 3; + * @return Whether the mergedInputDescriptor field is set. */ boolean hasMergedInputDescriptor(); /** * optional .EntityDescriptorProto merged_input_descriptor = 3; + * @return The mergedInputDescriptor. */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getMergedInputDescriptor(); /** @@ -2367,35 +2736,39 @@ public interface GroupInputSpecProtoOrBuilder * Protobuf type {@code GroupInputSpecProto} */ public static final class GroupInputSpecProto extends - com.google.protobuf.GeneratedMessage - implements GroupInputSpecProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:GroupInputSpecProto) + GroupInputSpecProtoOrBuilder { + private static final long serialVersionUID = 0L; // Use GroupInputSpecProto.newBuilder() to construct. - private GroupInputSpecProto(com.google.protobuf.GeneratedMessage.Builder builder) { + private GroupInputSpecProto(com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private GroupInputSpecProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final GroupInputSpecProto defaultInstance; - public static GroupInputSpecProto getDefaultInstance() { - return defaultInstance; + private GroupInputSpecProto() { + groupName_ = ""; + groupVertices_ = com.google.protobuf.LazyStringArrayList.EMPTY; } - public GroupInputSpecProto getDefaultInstanceForType() { - return defaultInstance; + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance( + UnusedPrivateParameter unused) { + return new GroupInputSpecProto(); } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private GroupInputSpecProto( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); @@ -2407,29 +2780,24 @@ private GroupInputSpecProto( case 0: done = true; break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } case 10: { + com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000001; - groupName_ = input.readBytes(); + groupName_ = bs; break; } case 18: { - if (!((mutable_bitField0_ & 0x00000002) == 0x00000002)) { + com.google.protobuf.ByteString bs = input.readBytes(); + if (!((mutable_bitField0_ & 0x00000002) != 0)) { groupVertices_ = new com.google.protobuf.LazyStringArrayList(); mutable_bitField0_ |= 0x00000002; } - groupVertices_.add(input.readBytes()); + groupVertices_.add(bs); break; } case 26: { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder subBuilder = null; - if (((bitField0_ & 0x00000002) == 0x00000002)) { + if (((bitField0_ & 0x00000002) != 0)) { subBuilder = mergedInputDescriptor_.toBuilder(); } mergedInputDescriptor_ = input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.PARSER, extensionRegistry); @@ -2440,16 +2808,23 @@ private GroupInputSpecProto( bitField0_ |= 0x00000002; break; } + default: { + if (!parseUnknownField( + input, unknownFields, extensionRegistry, tag)) { + done = true; + } + break; + } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { - if (((mutable_bitField0_ & 0x00000002) == 0x00000002)) { - groupVertices_ = new com.google.protobuf.UnmodifiableLazyStringList(groupVertices_); + if (((mutable_bitField0_ & 0x00000002) != 0)) { + groupVertices_ = groupVertices_.getUnmodifiableView(); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); @@ -2460,41 +2835,30 @@ private GroupInputSpecProto( return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GroupInputSpecProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GroupInputSpecProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public GroupInputSpecProto parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new GroupInputSpecProto(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - private int bitField0_; - // optional string group_name = 1; public static final int GROUP_NAME_FIELD_NUMBER = 1; - private java.lang.Object groupName_; + private volatile java.lang.Object groupName_; /** * optional string group_name = 1; + * @return Whether the groupName field is set. */ + @java.lang.Override public boolean hasGroupName() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional string group_name = 1; + * @return The groupName. */ + @java.lang.Override public java.lang.String getGroupName() { java.lang.Object ref = groupName_; if (ref instanceof java.lang.String) { @@ -2511,7 +2875,9 @@ public java.lang.String getGroupName() { } /** * optional string group_name = 1; + * @return The bytes for groupName. */ + @java.lang.Override public com.google.protobuf.ByteString getGroupNameBytes() { java.lang.Object ref = groupName_; @@ -2526,122 +2892,119 @@ public java.lang.String getGroupName() { } } - // repeated string group_vertices = 2; public static final int GROUP_VERTICES_FIELD_NUMBER = 2; private com.google.protobuf.LazyStringList groupVertices_; /** * repeated string group_vertices = 2; + * @return A list containing the groupVertices. */ - public java.util.List + public com.google.protobuf.ProtocolStringList getGroupVerticesList() { return groupVertices_; } /** * repeated string group_vertices = 2; + * @return The count of groupVertices. */ public int getGroupVerticesCount() { return groupVertices_.size(); } /** * repeated string group_vertices = 2; + * @param index The index of the element to return. + * @return The groupVertices at the given index. */ public java.lang.String getGroupVertices(int index) { return groupVertices_.get(index); } /** * repeated string group_vertices = 2; + * @param index The index of the value to return. + * @return The bytes of the groupVertices at the given index. */ public com.google.protobuf.ByteString getGroupVerticesBytes(int index) { return groupVertices_.getByteString(index); } - // optional .EntityDescriptorProto merged_input_descriptor = 3; public static final int MERGED_INPUT_DESCRIPTOR_FIELD_NUMBER = 3; private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto mergedInputDescriptor_; /** * optional .EntityDescriptorProto merged_input_descriptor = 3; + * @return Whether the mergedInputDescriptor field is set. */ + @java.lang.Override public boolean hasMergedInputDescriptor() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * optional .EntityDescriptorProto merged_input_descriptor = 3; + * @return The mergedInputDescriptor. */ + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getMergedInputDescriptor() { - return mergedInputDescriptor_; + return mergedInputDescriptor_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance() : mergedInputDescriptor_; } /** * optional .EntityDescriptorProto merged_input_descriptor = 3; */ + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder getMergedInputDescriptorOrBuilder() { - return mergedInputDescriptor_; + return mergedInputDescriptor_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance() : mergedInputDescriptor_; } - private void initFields() { - groupName_ = ""; - groupVertices_ = com.google.protobuf.LazyStringArrayList.EMPTY; - mergedInputDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance(); - } private byte memoizedIsInitialized = -1; + @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } + @java.lang.Override public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); - if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeBytes(1, getGroupNameBytes()); + if (((bitField0_ & 0x00000001) != 0)) { + com.google.protobuf.GeneratedMessageV3.writeString(output, 1, groupName_); } for (int i = 0; i < groupVertices_.size(); i++) { - output.writeBytes(2, groupVertices_.getByteString(i)); + com.google.protobuf.GeneratedMessageV3.writeString(output, 2, groupVertices_.getRaw(i)); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - output.writeMessage(3, mergedInputDescriptor_); + if (((bitField0_ & 0x00000002) != 0)) { + output.writeMessage(3, getMergedInputDescriptor()); } - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; + @java.lang.Override public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(1, getGroupNameBytes()); + if (((bitField0_ & 0x00000001) != 0)) { + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, groupName_); } { int dataSize = 0; for (int i = 0; i < groupVertices_.size(); i++) { - dataSize += com.google.protobuf.CodedOutputStream - .computeBytesSizeNoTag(groupVertices_.getByteString(i)); + dataSize += computeStringSizeNoTag(groupVertices_.getRaw(i)); } size += dataSize; size += 1 * getGroupVerticesList().size(); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { + if (((bitField0_ & 0x00000002) != 0)) { size += com.google.protobuf.CodedOutputStream - .computeMessageSize(3, mergedInputDescriptor_); + .computeMessageSize(3, getMergedInputDescriptor()); } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { @@ -2652,32 +3015,29 @@ public boolean equals(final java.lang.Object obj) { } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto) obj; - boolean result = true; - result = result && (hasGroupName() == other.hasGroupName()); + if (hasGroupName() != other.hasGroupName()) return false; if (hasGroupName()) { - result = result && getGroupName() - .equals(other.getGroupName()); + if (!getGroupName() + .equals(other.getGroupName())) return false; } - result = result && getGroupVerticesList() - .equals(other.getGroupVerticesList()); - result = result && (hasMergedInputDescriptor() == other.hasMergedInputDescriptor()); + if (!getGroupVerticesList() + .equals(other.getGroupVerticesList())) return false; + if (hasMergedInputDescriptor() != other.hasMergedInputDescriptor()) return false; if (hasMergedInputDescriptor()) { - result = result && getMergedInputDescriptor() - .equals(other.getMergedInputDescriptor()); + if (!getMergedInputDescriptor() + .equals(other.getMergedInputDescriptor())) return false; } - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; + if (!unknownFields.equals(other.unknownFields)) return false; + return true; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); + hash = (19 * hash) + getDescriptor().hashCode(); if (hasGroupName()) { hash = (37 * hash) + GROUP_NAME_FIELD_NUMBER; hash = (53 * hash) + getGroupName().hashCode(); @@ -2690,11 +3050,22 @@ public int hashCode() { hash = (37 * hash) + MERGED_INPUT_DESCRIPTOR_FIELD_NUMBER; hash = (53 * hash) + getMergedInputDescriptor().hashCode(); } - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto parseFrom( + java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto parseFrom( + java.nio.ByteBuffer data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { @@ -2718,46 +3089,59 @@ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.Gr } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } + @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -2765,14 +3149,16 @@ protected Builder newBuilderForType( * Protobuf type {@code GroupInputSpecProto} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder implements + // @@protoc_insertion_point(builder_implements:GroupInputSpecProto) + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProtoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GroupInputSpecProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GroupInputSpecProto_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -2785,19 +3171,17 @@ private Builder() { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { getMergedInputDescriptorFieldBuilder(); } } - private static Builder create() { - return new Builder(); - } - + @java.lang.Override public Builder clear() { super.clear(); groupName_ = ""; @@ -2805,7 +3189,7 @@ public Builder clear() { groupVertices_ = com.google.protobuf.LazyStringArrayList.EMPTY; bitField0_ = (bitField0_ & ~0x00000002); if (mergedInputDescriptorBuilder_ == null) { - mergedInputDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance(); + mergedInputDescriptor_ = null; } else { mergedInputDescriptorBuilder_.clear(); } @@ -2813,19 +3197,18 @@ public Builder clear() { return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - + @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GroupInputSpecProto_descriptor; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.getDefaultInstance(); } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto result = buildPartial(); if (!result.isInitialized()) { @@ -2834,33 +3217,66 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInpu return result; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { + if (((from_bitField0_ & 0x00000001) != 0)) { to_bitField0_ |= 0x00000001; } result.groupName_ = groupName_; - if (((bitField0_ & 0x00000002) == 0x00000002)) { - groupVertices_ = new com.google.protobuf.UnmodifiableLazyStringList( - groupVertices_); + if (((bitField0_ & 0x00000002) != 0)) { + groupVertices_ = groupVertices_.getUnmodifiableView(); bitField0_ = (bitField0_ & ~0x00000002); } result.groupVertices_ = groupVertices_; - if (((from_bitField0_ & 0x00000004) == 0x00000004)) { + if (((from_bitField0_ & 0x00000004) != 0)) { + if (mergedInputDescriptorBuilder_ == null) { + result.mergedInputDescriptor_ = mergedInputDescriptor_; + } else { + result.mergedInputDescriptor_ = mergedInputDescriptorBuilder_.build(); + } to_bitField0_ |= 0x00000002; } - if (mergedInputDescriptorBuilder_ == null) { - result.mergedInputDescriptor_ = mergedInputDescriptor_; - } else { - result.mergedInputDescriptor_ = mergedInputDescriptorBuilder_.build(); - } result.bitField0_ = to_bitField0_; onBuilt(); return result; } + @java.lang.Override + public Builder clone() { + return super.clone(); + } + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.setField(field, value); + } + @java.lang.Override + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + @java.lang.Override + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.addRepeatedField(field, value); + } + @java.lang.Override public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto)other); @@ -2890,14 +3306,17 @@ public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtoc if (other.hasMergedInputDescriptor()) { mergeMergedInputDescriptor(other.getMergedInputDescriptor()); } - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } + @java.lang.Override public final boolean isInitialized() { return true; } + @java.lang.Override public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) @@ -2907,7 +3326,7 @@ public Builder mergeFrom( parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -2917,23 +3336,27 @@ public Builder mergeFrom( } private int bitField0_; - // optional string group_name = 1; private java.lang.Object groupName_ = ""; /** * optional string group_name = 1; + * @return Whether the groupName field is set. */ public boolean hasGroupName() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional string group_name = 1; + * @return The groupName. */ public java.lang.String getGroupName() { java.lang.Object ref = groupName_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - groupName_ = s; + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + groupName_ = s; + } return s; } else { return (java.lang.String) ref; @@ -2941,6 +3364,7 @@ public java.lang.String getGroupName() { } /** * optional string group_name = 1; + * @return The bytes for groupName. */ public com.google.protobuf.ByteString getGroupNameBytes() { @@ -2957,6 +3381,8 @@ public java.lang.String getGroupName() { } /** * optional string group_name = 1; + * @param value The groupName to set. + * @return This builder for chaining. */ public Builder setGroupName( java.lang.String value) { @@ -2970,6 +3396,7 @@ public Builder setGroupName( } /** * optional string group_name = 1; + * @return This builder for chaining. */ public Builder clearGroupName() { bitField0_ = (bitField0_ & ~0x00000001); @@ -2979,6 +3406,8 @@ public Builder clearGroupName() { } /** * optional string group_name = 1; + * @param value The bytes for groupName to set. + * @return This builder for chaining. */ public Builder setGroupNameBytes( com.google.protobuf.ByteString value) { @@ -2991,35 +3420,40 @@ public Builder setGroupNameBytes( return this; } - // repeated string group_vertices = 2; private com.google.protobuf.LazyStringList groupVertices_ = com.google.protobuf.LazyStringArrayList.EMPTY; private void ensureGroupVerticesIsMutable() { - if (!((bitField0_ & 0x00000002) == 0x00000002)) { + if (!((bitField0_ & 0x00000002) != 0)) { groupVertices_ = new com.google.protobuf.LazyStringArrayList(groupVertices_); bitField0_ |= 0x00000002; } } /** * repeated string group_vertices = 2; + * @return A list containing the groupVertices. */ - public java.util.List + public com.google.protobuf.ProtocolStringList getGroupVerticesList() { - return java.util.Collections.unmodifiableList(groupVertices_); + return groupVertices_.getUnmodifiableView(); } /** * repeated string group_vertices = 2; + * @return The count of groupVertices. */ public int getGroupVerticesCount() { return groupVertices_.size(); } /** * repeated string group_vertices = 2; + * @param index The index of the element to return. + * @return The groupVertices at the given index. */ public java.lang.String getGroupVertices(int index) { return groupVertices_.get(index); } /** * repeated string group_vertices = 2; + * @param index The index of the value to return. + * @return The bytes of the groupVertices at the given index. */ public com.google.protobuf.ByteString getGroupVerticesBytes(int index) { @@ -3027,6 +3461,9 @@ public java.lang.String getGroupVertices(int index) { } /** * repeated string group_vertices = 2; + * @param index The index to set the value at. + * @param value The groupVertices to set. + * @return This builder for chaining. */ public Builder setGroupVertices( int index, java.lang.String value) { @@ -3040,6 +3477,8 @@ public Builder setGroupVertices( } /** * repeated string group_vertices = 2; + * @param value The groupVertices to add. + * @return This builder for chaining. */ public Builder addGroupVertices( java.lang.String value) { @@ -3053,16 +3492,20 @@ public Builder addGroupVertices( } /** * repeated string group_vertices = 2; + * @param values The groupVertices to add. + * @return This builder for chaining. */ public Builder addAllGroupVertices( java.lang.Iterable values) { ensureGroupVerticesIsMutable(); - super.addAll(values, groupVertices_); + com.google.protobuf.AbstractMessageLite.Builder.addAll( + values, groupVertices_); onChanged(); return this; } /** * repeated string group_vertices = 2; + * @return This builder for chaining. */ public Builder clearGroupVertices() { groupVertices_ = com.google.protobuf.LazyStringArrayList.EMPTY; @@ -3072,6 +3515,8 @@ public Builder clearGroupVertices() { } /** * repeated string group_vertices = 2; + * @param value The bytes of the groupVertices to add. + * @return This builder for chaining. */ public Builder addGroupVerticesBytes( com.google.protobuf.ByteString value) { @@ -3084,22 +3529,23 @@ public Builder addGroupVerticesBytes( return this; } - // optional .EntityDescriptorProto merged_input_descriptor = 3; - private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto mergedInputDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance(); - private com.google.protobuf.SingleFieldBuilder< + private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto mergedInputDescriptor_; + private com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder> mergedInputDescriptorBuilder_; /** * optional .EntityDescriptorProto merged_input_descriptor = 3; + * @return Whether the mergedInputDescriptor field is set. */ public boolean hasMergedInputDescriptor() { - return ((bitField0_ & 0x00000004) == 0x00000004); + return ((bitField0_ & 0x00000004) != 0); } /** * optional .EntityDescriptorProto merged_input_descriptor = 3; + * @return The mergedInputDescriptor. */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getMergedInputDescriptor() { if (mergedInputDescriptorBuilder_ == null) { - return mergedInputDescriptor_; + return mergedInputDescriptor_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance() : mergedInputDescriptor_; } else { return mergedInputDescriptorBuilder_.getMessage(); } @@ -3139,7 +3585,8 @@ public Builder setMergedInputDescriptor( */ public Builder mergeMergedInputDescriptor(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto value) { if (mergedInputDescriptorBuilder_ == null) { - if (((bitField0_ & 0x00000004) == 0x00000004) && + if (((bitField0_ & 0x00000004) != 0) && + mergedInputDescriptor_ != null && mergedInputDescriptor_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance()) { mergedInputDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.newBuilder(mergedInputDescriptor_).mergeFrom(value).buildPartial(); @@ -3158,7 +3605,7 @@ public Builder mergeMergedInputDescriptor(org.apache.hadoop.hive.llap.daemon.rpc */ public Builder clearMergedInputDescriptor() { if (mergedInputDescriptorBuilder_ == null) { - mergedInputDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance(); + mergedInputDescriptor_ = null; onChanged(); } else { mergedInputDescriptorBuilder_.clear(); @@ -3181,72 +3628,119 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDes if (mergedInputDescriptorBuilder_ != null) { return mergedInputDescriptorBuilder_.getMessageOrBuilder(); } else { - return mergedInputDescriptor_; + return mergedInputDescriptor_ == null ? + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance() : mergedInputDescriptor_; } } /** * optional .EntityDescriptorProto merged_input_descriptor = 3; */ - private com.google.protobuf.SingleFieldBuilder< + private com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder> getMergedInputDescriptorFieldBuilder() { if (mergedInputDescriptorBuilder_ == null) { - mergedInputDescriptorBuilder_ = new com.google.protobuf.SingleFieldBuilder< + mergedInputDescriptorBuilder_ = new com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder>( - mergedInputDescriptor_, + getMergedInputDescriptor(), getParentForChildren(), isClean()); mergedInputDescriptor_ = null; } return mergedInputDescriptorBuilder_; } + @java.lang.Override + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:GroupInputSpecProto) } + // @@protoc_insertion_point(class_scope:GroupInputSpecProto) + private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto DEFAULT_INSTANCE; static { - defaultInstance = new GroupInputSpecProto(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto(); + } + + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser + PARSER = new com.google.protobuf.AbstractParser() { + @java.lang.Override + public GroupInputSpecProto parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new GroupInputSpecProto(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto getDefaultInstanceForType() { + return DEFAULT_INSTANCE; } - // @@protoc_insertion_point(class_scope:GroupInputSpecProto) } - public interface SignableVertexSpecOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface SignableVertexSpecOrBuilder extends + // @@protoc_insertion_point(interface_extends:SignableVertexSpec) + com.google.protobuf.MessageOrBuilder { - // optional string user = 1; /** * optional string user = 1; + * @return Whether the user field is set. */ boolean hasUser(); /** * optional string user = 1; + * @return The user. */ java.lang.String getUser(); /** * optional string user = 1; + * @return The bytes for user. */ com.google.protobuf.ByteString getUserBytes(); - // optional int64 signatureKeyId = 2; /** * optional int64 signatureKeyId = 2; + * @return Whether the signatureKeyId field is set. */ boolean hasSignatureKeyId(); /** * optional int64 signatureKeyId = 2; + * @return The signatureKeyId. */ long getSignatureKeyId(); - // optional .QueryIdentifierProto query_identifier = 3; /** * optional .QueryIdentifierProto query_identifier = 3; + * @return Whether the queryIdentifier field is set. */ boolean hasQueryIdentifier(); /** * optional .QueryIdentifierProto query_identifier = 3; + * @return The queryIdentifier. */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto getQueryIdentifier(); /** @@ -3254,107 +3748,117 @@ public interface SignableVertexSpecOrBuilder */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder getQueryIdentifierOrBuilder(); - // optional string hive_query_id = 4; /** * optional string hive_query_id = 4; + * @return Whether the hiveQueryId field is set. */ boolean hasHiveQueryId(); /** * optional string hive_query_id = 4; + * @return The hiveQueryId. */ java.lang.String getHiveQueryId(); /** * optional string hive_query_id = 4; + * @return The bytes for hiveQueryId. */ com.google.protobuf.ByteString getHiveQueryIdBytes(); - // optional string dag_name = 5; /** - * optional string dag_name = 5; - * *
      * Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here.
      * 
+ * + * optional string dag_name = 5; + * @return Whether the dagName field is set. */ boolean hasDagName(); /** - * optional string dag_name = 5; - * *
      * Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here.
      * 
+ * + * optional string dag_name = 5; + * @return The dagName. */ java.lang.String getDagName(); /** - * optional string dag_name = 5; - * *
      * Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here.
      * 
+ * + * optional string dag_name = 5; + * @return The bytes for dagName. */ com.google.protobuf.ByteString getDagNameBytes(); - // optional string vertex_name = 6; /** * optional string vertex_name = 6; + * @return Whether the vertexName field is set. */ boolean hasVertexName(); /** * optional string vertex_name = 6; + * @return The vertexName. */ java.lang.String getVertexName(); /** * optional string vertex_name = 6; + * @return The bytes for vertexName. */ com.google.protobuf.ByteString getVertexNameBytes(); - // optional int32 vertex_index = 7; /** * optional int32 vertex_index = 7; + * @return Whether the vertexIndex field is set. */ boolean hasVertexIndex(); /** * optional int32 vertex_index = 7; + * @return The vertexIndex. */ int getVertexIndex(); - // optional string token_identifier = 8; /** - * optional string token_identifier = 8; - * *
      * The core vertex stuff 
      * 
+ * + * optional string token_identifier = 8; + * @return Whether the tokenIdentifier field is set. */ boolean hasTokenIdentifier(); /** - * optional string token_identifier = 8; - * *
      * The core vertex stuff 
      * 
+ * + * optional string token_identifier = 8; + * @return The tokenIdentifier. */ java.lang.String getTokenIdentifier(); /** - * optional string token_identifier = 8; - * *
      * The core vertex stuff 
      * 
+ * + * optional string token_identifier = 8; + * @return The bytes for tokenIdentifier. */ com.google.protobuf.ByteString getTokenIdentifierBytes(); - // optional .EntityDescriptorProto processor_descriptor = 9; /** * optional .EntityDescriptorProto processor_descriptor = 9; + * @return Whether the processorDescriptor field is set. */ boolean hasProcessorDescriptor(); /** * optional .EntityDescriptorProto processor_descriptor = 9; + * @return The processorDescriptor. */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getProcessorDescriptor(); /** @@ -3362,7 +3866,6 @@ public interface SignableVertexSpecOrBuilder */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder getProcessorDescriptorOrBuilder(); - // repeated .IOSpecProto input_specs = 10; /** * repeated .IOSpecProto input_specs = 10; */ @@ -3387,7 +3890,6 @@ public interface SignableVertexSpecOrBuilder org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder getInputSpecsOrBuilder( int index); - // repeated .IOSpecProto output_specs = 11; /** * repeated .IOSpecProto output_specs = 11; */ @@ -3412,7 +3914,6 @@ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBui org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder getOutputSpecsOrBuilder( int index); - // repeated .GroupInputSpecProto grouped_input_specs = 12; /** * repeated .GroupInputSpecProto grouped_input_specs = 12; */ @@ -3437,71 +3938,83 @@ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBui org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProtoOrBuilder getGroupedInputSpecsOrBuilder( int index); - // optional int32 vertex_parallelism = 13; /** - * optional int32 vertex_parallelism = 13; - * *
      * An internal field required for Tez.
      * 
+ * + * optional int32 vertex_parallelism = 13; + * @return Whether the vertexParallelism field is set. */ boolean hasVertexParallelism(); /** - * optional int32 vertex_parallelism = 13; - * *
      * An internal field required for Tez.
      * 
+ * + * optional int32 vertex_parallelism = 13; + * @return The vertexParallelism. */ int getVertexParallelism(); - // optional bool is_external_submission = 14 [default = false]; /** * optional bool is_external_submission = 14 [default = false]; + * @return Whether the isExternalSubmission field is set. */ boolean hasIsExternalSubmission(); /** * optional bool is_external_submission = 14 [default = false]; + * @return The isExternalSubmission. */ boolean getIsExternalSubmission(); } /** - * Protobuf type {@code SignableVertexSpec} - * *
    * The part of SubmitWork that can be signed 
    * 
+ * + * Protobuf type {@code SignableVertexSpec} */ public static final class SignableVertexSpec extends - com.google.protobuf.GeneratedMessage - implements SignableVertexSpecOrBuilder { + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:SignableVertexSpec) + SignableVertexSpecOrBuilder { + private static final long serialVersionUID = 0L; // Use SignableVertexSpec.newBuilder() to construct. - private SignableVertexSpec(com.google.protobuf.GeneratedMessage.Builder builder) { + private SignableVertexSpec(com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private SignableVertexSpec(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final SignableVertexSpec defaultInstance; - public static SignableVertexSpec getDefaultInstance() { - return defaultInstance; + private SignableVertexSpec() { + user_ = ""; + hiveQueryId_ = ""; + dagName_ = ""; + vertexName_ = ""; + tokenIdentifier_ = ""; + inputSpecs_ = java.util.Collections.emptyList(); + outputSpecs_ = java.util.Collections.emptyList(); + groupedInputSpecs_ = java.util.Collections.emptyList(); } - public SignableVertexSpec getDefaultInstanceForType() { - return defaultInstance; + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance( + UnusedPrivateParameter unused) { + return new SignableVertexSpec(); } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private SignableVertexSpec( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); @@ -3513,16 +4026,10 @@ private SignableVertexSpec( case 0: done = true; break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } case 10: { + com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000001; - user_ = input.readBytes(); + user_ = bs; break; } case 16: { @@ -3532,7 +4039,7 @@ private SignableVertexSpec( } case 26: { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder subBuilder = null; - if (((bitField0_ & 0x00000004) == 0x00000004)) { + if (((bitField0_ & 0x00000004) != 0)) { subBuilder = queryIdentifier_.toBuilder(); } queryIdentifier_ = input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.PARSER, extensionRegistry); @@ -3544,18 +4051,21 @@ private SignableVertexSpec( break; } case 34: { + com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000008; - hiveQueryId_ = input.readBytes(); + hiveQueryId_ = bs; break; } case 42: { + com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000010; - dagName_ = input.readBytes(); + dagName_ = bs; break; } case 50: { + com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000020; - vertexName_ = input.readBytes(); + vertexName_ = bs; break; } case 56: { @@ -3564,13 +4074,14 @@ private SignableVertexSpec( break; } case 66: { + com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000080; - tokenIdentifier_ = input.readBytes(); + tokenIdentifier_ = bs; break; } case 74: { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder subBuilder = null; - if (((bitField0_ & 0x00000100) == 0x00000100)) { + if (((bitField0_ & 0x00000100) != 0)) { subBuilder = processorDescriptor_.toBuilder(); } processorDescriptor_ = input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.PARSER, extensionRegistry); @@ -3582,27 +4093,30 @@ private SignableVertexSpec( break; } case 82: { - if (!((mutable_bitField0_ & 0x00000200) == 0x00000200)) { + if (!((mutable_bitField0_ & 0x00000200) != 0)) { inputSpecs_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000200; } - inputSpecs_.add(input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.PARSER, extensionRegistry)); + inputSpecs_.add( + input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.PARSER, extensionRegistry)); break; } case 90: { - if (!((mutable_bitField0_ & 0x00000400) == 0x00000400)) { + if (!((mutable_bitField0_ & 0x00000400) != 0)) { outputSpecs_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000400; } - outputSpecs_.add(input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.PARSER, extensionRegistry)); + outputSpecs_.add( + input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.PARSER, extensionRegistry)); break; } case 98: { - if (!((mutable_bitField0_ & 0x00000800) == 0x00000800)) { + if (!((mutable_bitField0_ & 0x00000800) != 0)) { groupedInputSpecs_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000800; } - groupedInputSpecs_.add(input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.PARSER, extensionRegistry)); + groupedInputSpecs_.add( + input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.PARSER, extensionRegistry)); break; } case 104: { @@ -3615,21 +4129,28 @@ private SignableVertexSpec( isExternalSubmission_ = input.readBool(); break; } + default: { + if (!parseUnknownField( + input, unknownFields, extensionRegistry, tag)) { + done = true; + } + break; + } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { - if (((mutable_bitField0_ & 0x00000200) == 0x00000200)) { + if (((mutable_bitField0_ & 0x00000200) != 0)) { inputSpecs_ = java.util.Collections.unmodifiableList(inputSpecs_); } - if (((mutable_bitField0_ & 0x00000400) == 0x00000400)) { + if (((mutable_bitField0_ & 0x00000400) != 0)) { outputSpecs_ = java.util.Collections.unmodifiableList(outputSpecs_); } - if (((mutable_bitField0_ & 0x00000800) == 0x00000800)) { + if (((mutable_bitField0_ & 0x00000800) != 0)) { groupedInputSpecs_ = java.util.Collections.unmodifiableList(groupedInputSpecs_); } this.unknownFields = unknownFields.build(); @@ -3641,41 +4162,30 @@ private SignableVertexSpec( return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SignableVertexSpec_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SignableVertexSpec_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public SignableVertexSpec parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new SignableVertexSpec(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - private int bitField0_; - // optional string user = 1; public static final int USER_FIELD_NUMBER = 1; - private java.lang.Object user_; + private volatile java.lang.Object user_; /** * optional string user = 1; + * @return Whether the user field is set. */ + @java.lang.Override public boolean hasUser() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional string user = 1; + * @return The user. */ + @java.lang.Override public java.lang.String getUser() { java.lang.Object ref = user_; if (ref instanceof java.lang.String) { @@ -3692,7 +4202,9 @@ public java.lang.String getUser() { } /** * optional string user = 1; + * @return The bytes for user. */ + @java.lang.Override public com.google.protobuf.ByteString getUserBytes() { java.lang.Object ref = user_; @@ -3707,56 +4219,66 @@ public java.lang.String getUser() { } } - // optional int64 signatureKeyId = 2; public static final int SIGNATUREKEYID_FIELD_NUMBER = 2; private long signatureKeyId_; /** * optional int64 signatureKeyId = 2; + * @return Whether the signatureKeyId field is set. */ + @java.lang.Override public boolean hasSignatureKeyId() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * optional int64 signatureKeyId = 2; + * @return The signatureKeyId. */ + @java.lang.Override public long getSignatureKeyId() { return signatureKeyId_; } - // optional .QueryIdentifierProto query_identifier = 3; public static final int QUERY_IDENTIFIER_FIELD_NUMBER = 3; private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto queryIdentifier_; /** * optional .QueryIdentifierProto query_identifier = 3; + * @return Whether the queryIdentifier field is set. */ + @java.lang.Override public boolean hasQueryIdentifier() { - return ((bitField0_ & 0x00000004) == 0x00000004); + return ((bitField0_ & 0x00000004) != 0); } /** * optional .QueryIdentifierProto query_identifier = 3; + * @return The queryIdentifier. */ + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto getQueryIdentifier() { - return queryIdentifier_; + return queryIdentifier_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance() : queryIdentifier_; } /** * optional .QueryIdentifierProto query_identifier = 3; */ + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder getQueryIdentifierOrBuilder() { - return queryIdentifier_; + return queryIdentifier_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance() : queryIdentifier_; } - // optional string hive_query_id = 4; public static final int HIVE_QUERY_ID_FIELD_NUMBER = 4; - private java.lang.Object hiveQueryId_; + private volatile java.lang.Object hiveQueryId_; /** * optional string hive_query_id = 4; + * @return Whether the hiveQueryId field is set. */ + @java.lang.Override public boolean hasHiveQueryId() { - return ((bitField0_ & 0x00000008) == 0x00000008); + return ((bitField0_ & 0x00000008) != 0); } /** * optional string hive_query_id = 4; + * @return The hiveQueryId. */ + @java.lang.Override public java.lang.String getHiveQueryId() { java.lang.Object ref = hiveQueryId_; if (ref instanceof java.lang.String) { @@ -3773,7 +4295,9 @@ public java.lang.String getHiveQueryId() { } /** * optional string hive_query_id = 4; + * @return The bytes for hiveQueryId. */ + @java.lang.Override public com.google.protobuf.ByteString getHiveQueryIdBytes() { java.lang.Object ref = hiveQueryId_; @@ -3788,26 +4312,29 @@ public java.lang.String getHiveQueryId() { } } - // optional string dag_name = 5; public static final int DAG_NAME_FIELD_NUMBER = 5; - private java.lang.Object dagName_; + private volatile java.lang.Object dagName_; /** - * optional string dag_name = 5; - * *
      * Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here.
      * 
+ * + * optional string dag_name = 5; + * @return Whether the dagName field is set. */ + @java.lang.Override public boolean hasDagName() { - return ((bitField0_ & 0x00000010) == 0x00000010); + return ((bitField0_ & 0x00000010) != 0); } /** - * optional string dag_name = 5; - * *
      * Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here.
      * 
+ * + * optional string dag_name = 5; + * @return The dagName. */ + @java.lang.Override public java.lang.String getDagName() { java.lang.Object ref = dagName_; if (ref instanceof java.lang.String) { @@ -3823,12 +4350,14 @@ public java.lang.String getDagName() { } } /** - * optional string dag_name = 5; - * *
      * Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here.
      * 
+ * + * optional string dag_name = 5; + * @return The bytes for dagName. */ + @java.lang.Override public com.google.protobuf.ByteString getDagNameBytes() { java.lang.Object ref = dagName_; @@ -3843,18 +4372,21 @@ public java.lang.String getDagName() { } } - // optional string vertex_name = 6; public static final int VERTEX_NAME_FIELD_NUMBER = 6; - private java.lang.Object vertexName_; + private volatile java.lang.Object vertexName_; /** * optional string vertex_name = 6; + * @return Whether the vertexName field is set. */ + @java.lang.Override public boolean hasVertexName() { - return ((bitField0_ & 0x00000020) == 0x00000020); + return ((bitField0_ & 0x00000020) != 0); } /** * optional string vertex_name = 6; + * @return The vertexName. */ + @java.lang.Override public java.lang.String getVertexName() { java.lang.Object ref = vertexName_; if (ref instanceof java.lang.String) { @@ -3871,7 +4403,9 @@ public java.lang.String getVertexName() { } /** * optional string vertex_name = 6; + * @return The bytes for vertexName. */ + @java.lang.Override public com.google.protobuf.ByteString getVertexNameBytes() { java.lang.Object ref = vertexName_; @@ -3886,42 +4420,48 @@ public java.lang.String getVertexName() { } } - // optional int32 vertex_index = 7; public static final int VERTEX_INDEX_FIELD_NUMBER = 7; private int vertexIndex_; /** * optional int32 vertex_index = 7; + * @return Whether the vertexIndex field is set. */ + @java.lang.Override public boolean hasVertexIndex() { - return ((bitField0_ & 0x00000040) == 0x00000040); + return ((bitField0_ & 0x00000040) != 0); } /** * optional int32 vertex_index = 7; + * @return The vertexIndex. */ + @java.lang.Override public int getVertexIndex() { return vertexIndex_; } - // optional string token_identifier = 8; public static final int TOKEN_IDENTIFIER_FIELD_NUMBER = 8; - private java.lang.Object tokenIdentifier_; + private volatile java.lang.Object tokenIdentifier_; /** - * optional string token_identifier = 8; - * *
      * The core vertex stuff 
      * 
+ * + * optional string token_identifier = 8; + * @return Whether the tokenIdentifier field is set. */ + @java.lang.Override public boolean hasTokenIdentifier() { - return ((bitField0_ & 0x00000080) == 0x00000080); + return ((bitField0_ & 0x00000080) != 0); } /** - * optional string token_identifier = 8; - * *
      * The core vertex stuff 
      * 
+ * + * optional string token_identifier = 8; + * @return The tokenIdentifier. */ + @java.lang.Override public java.lang.String getTokenIdentifier() { java.lang.Object ref = tokenIdentifier_; if (ref instanceof java.lang.String) { @@ -3937,12 +4477,14 @@ public java.lang.String getTokenIdentifier() { } } /** - * optional string token_identifier = 8; - * *
      * The core vertex stuff 
      * 
+ * + * optional string token_identifier = 8; + * @return The bytes for tokenIdentifier. */ + @java.lang.Override public com.google.protobuf.ByteString getTokenIdentifierBytes() { java.lang.Object ref = tokenIdentifier_; @@ -3957,40 +4499,45 @@ public java.lang.String getTokenIdentifier() { } } - // optional .EntityDescriptorProto processor_descriptor = 9; public static final int PROCESSOR_DESCRIPTOR_FIELD_NUMBER = 9; private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto processorDescriptor_; /** * optional .EntityDescriptorProto processor_descriptor = 9; + * @return Whether the processorDescriptor field is set. */ + @java.lang.Override public boolean hasProcessorDescriptor() { - return ((bitField0_ & 0x00000100) == 0x00000100); + return ((bitField0_ & 0x00000100) != 0); } /** * optional .EntityDescriptorProto processor_descriptor = 9; + * @return The processorDescriptor. */ + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getProcessorDescriptor() { - return processorDescriptor_; + return processorDescriptor_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance() : processorDescriptor_; } /** * optional .EntityDescriptorProto processor_descriptor = 9; */ + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder getProcessorDescriptorOrBuilder() { - return processorDescriptor_; + return processorDescriptor_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance() : processorDescriptor_; } - // repeated .IOSpecProto input_specs = 10; public static final int INPUT_SPECS_FIELD_NUMBER = 10; private java.util.List inputSpecs_; /** * repeated .IOSpecProto input_specs = 10; */ + @java.lang.Override public java.util.List getInputSpecsList() { return inputSpecs_; } /** * repeated .IOSpecProto input_specs = 10; */ + @java.lang.Override public java.util.List getInputSpecsOrBuilderList() { return inputSpecs_; @@ -3998,35 +4545,39 @@ public java.util.Listrepeated .IOSpecProto input_specs = 10; */ + @java.lang.Override public int getInputSpecsCount() { return inputSpecs_.size(); } /** * repeated .IOSpecProto input_specs = 10; */ + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto getInputSpecs(int index) { return inputSpecs_.get(index); } /** * repeated .IOSpecProto input_specs = 10; */ + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder getInputSpecsOrBuilder( int index) { return inputSpecs_.get(index); } - // repeated .IOSpecProto output_specs = 11; public static final int OUTPUT_SPECS_FIELD_NUMBER = 11; private java.util.List outputSpecs_; /** * repeated .IOSpecProto output_specs = 11; */ + @java.lang.Override public java.util.List getOutputSpecsList() { return outputSpecs_; } /** * repeated .IOSpecProto output_specs = 11; */ + @java.lang.Override public java.util.List getOutputSpecsOrBuilderList() { return outputSpecs_; @@ -4034,35 +4585,39 @@ public java.util.Listrepeated .IOSpecProto output_specs = 11; */ + @java.lang.Override public int getOutputSpecsCount() { return outputSpecs_.size(); } /** * repeated .IOSpecProto output_specs = 11; */ + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto getOutputSpecs(int index) { return outputSpecs_.get(index); } /** * repeated .IOSpecProto output_specs = 11; */ + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder getOutputSpecsOrBuilder( int index) { return outputSpecs_.get(index); } - // repeated .GroupInputSpecProto grouped_input_specs = 12; public static final int GROUPED_INPUT_SPECS_FIELD_NUMBER = 12; private java.util.List groupedInputSpecs_; /** * repeated .GroupInputSpecProto grouped_input_specs = 12; */ + @java.lang.Override public java.util.List getGroupedInputSpecsList() { return groupedInputSpecs_; } /** * repeated .GroupInputSpecProto grouped_input_specs = 12; */ + @java.lang.Override public java.util.List getGroupedInputSpecsOrBuilderList() { return groupedInputSpecs_; @@ -4070,117 +4625,112 @@ public java.util.Listrepeated .GroupInputSpecProto grouped_input_specs = 12; */ + @java.lang.Override public int getGroupedInputSpecsCount() { return groupedInputSpecs_.size(); } /** * repeated .GroupInputSpecProto grouped_input_specs = 12; */ + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto getGroupedInputSpecs(int index) { return groupedInputSpecs_.get(index); } /** * repeated .GroupInputSpecProto grouped_input_specs = 12; */ + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProtoOrBuilder getGroupedInputSpecsOrBuilder( int index) { return groupedInputSpecs_.get(index); } - // optional int32 vertex_parallelism = 13; public static final int VERTEX_PARALLELISM_FIELD_NUMBER = 13; private int vertexParallelism_; /** - * optional int32 vertex_parallelism = 13; - * *
      * An internal field required for Tez.
      * 
+ * + * optional int32 vertex_parallelism = 13; + * @return Whether the vertexParallelism field is set. */ + @java.lang.Override public boolean hasVertexParallelism() { - return ((bitField0_ & 0x00000200) == 0x00000200); + return ((bitField0_ & 0x00000200) != 0); } /** - * optional int32 vertex_parallelism = 13; - * *
      * An internal field required for Tez.
      * 
+ * + * optional int32 vertex_parallelism = 13; + * @return The vertexParallelism. */ + @java.lang.Override public int getVertexParallelism() { return vertexParallelism_; } - // optional bool is_external_submission = 14 [default = false]; public static final int IS_EXTERNAL_SUBMISSION_FIELD_NUMBER = 14; private boolean isExternalSubmission_; /** * optional bool is_external_submission = 14 [default = false]; + * @return Whether the isExternalSubmission field is set. */ + @java.lang.Override public boolean hasIsExternalSubmission() { - return ((bitField0_ & 0x00000400) == 0x00000400); + return ((bitField0_ & 0x00000400) != 0); } /** * optional bool is_external_submission = 14 [default = false]; + * @return The isExternalSubmission. */ + @java.lang.Override public boolean getIsExternalSubmission() { return isExternalSubmission_; } - private void initFields() { - user_ = ""; - signatureKeyId_ = 0L; - queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance(); - hiveQueryId_ = ""; - dagName_ = ""; - vertexName_ = ""; - vertexIndex_ = 0; - tokenIdentifier_ = ""; - processorDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance(); - inputSpecs_ = java.util.Collections.emptyList(); - outputSpecs_ = java.util.Collections.emptyList(); - groupedInputSpecs_ = java.util.Collections.emptyList(); - vertexParallelism_ = 0; - isExternalSubmission_ = false; - } private byte memoizedIsInitialized = -1; + @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } + @java.lang.Override public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); - if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeBytes(1, getUserBytes()); + if (((bitField0_ & 0x00000001) != 0)) { + com.google.protobuf.GeneratedMessageV3.writeString(output, 1, user_); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { + if (((bitField0_ & 0x00000002) != 0)) { output.writeInt64(2, signatureKeyId_); } - if (((bitField0_ & 0x00000004) == 0x00000004)) { - output.writeMessage(3, queryIdentifier_); + if (((bitField0_ & 0x00000004) != 0)) { + output.writeMessage(3, getQueryIdentifier()); } - if (((bitField0_ & 0x00000008) == 0x00000008)) { - output.writeBytes(4, getHiveQueryIdBytes()); + if (((bitField0_ & 0x00000008) != 0)) { + com.google.protobuf.GeneratedMessageV3.writeString(output, 4, hiveQueryId_); } - if (((bitField0_ & 0x00000010) == 0x00000010)) { - output.writeBytes(5, getDagNameBytes()); + if (((bitField0_ & 0x00000010) != 0)) { + com.google.protobuf.GeneratedMessageV3.writeString(output, 5, dagName_); } - if (((bitField0_ & 0x00000020) == 0x00000020)) { - output.writeBytes(6, getVertexNameBytes()); + if (((bitField0_ & 0x00000020) != 0)) { + com.google.protobuf.GeneratedMessageV3.writeString(output, 6, vertexName_); } - if (((bitField0_ & 0x00000040) == 0x00000040)) { + if (((bitField0_ & 0x00000040) != 0)) { output.writeInt32(7, vertexIndex_); } - if (((bitField0_ & 0x00000080) == 0x00000080)) { - output.writeBytes(8, getTokenIdentifierBytes()); + if (((bitField0_ & 0x00000080) != 0)) { + com.google.protobuf.GeneratedMessageV3.writeString(output, 8, tokenIdentifier_); } - if (((bitField0_ & 0x00000100) == 0x00000100)) { - output.writeMessage(9, processorDescriptor_); + if (((bitField0_ & 0x00000100) != 0)) { + output.writeMessage(9, getProcessorDescriptor()); } for (int i = 0; i < inputSpecs_.size(); i++) { output.writeMessage(10, inputSpecs_.get(i)); @@ -4191,56 +4741,51 @@ public void writeTo(com.google.protobuf.CodedOutputStream output) for (int i = 0; i < groupedInputSpecs_.size(); i++) { output.writeMessage(12, groupedInputSpecs_.get(i)); } - if (((bitField0_ & 0x00000200) == 0x00000200)) { + if (((bitField0_ & 0x00000200) != 0)) { output.writeInt32(13, vertexParallelism_); } - if (((bitField0_ & 0x00000400) == 0x00000400)) { + if (((bitField0_ & 0x00000400) != 0)) { output.writeBool(14, isExternalSubmission_); } - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; + @java.lang.Override public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(1, getUserBytes()); + if (((bitField0_ & 0x00000001) != 0)) { + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, user_); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { + if (((bitField0_ & 0x00000002) != 0)) { size += com.google.protobuf.CodedOutputStream .computeInt64Size(2, signatureKeyId_); } - if (((bitField0_ & 0x00000004) == 0x00000004)) { + if (((bitField0_ & 0x00000004) != 0)) { size += com.google.protobuf.CodedOutputStream - .computeMessageSize(3, queryIdentifier_); + .computeMessageSize(3, getQueryIdentifier()); } - if (((bitField0_ & 0x00000008) == 0x00000008)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(4, getHiveQueryIdBytes()); + if (((bitField0_ & 0x00000008) != 0)) { + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(4, hiveQueryId_); } - if (((bitField0_ & 0x00000010) == 0x00000010)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(5, getDagNameBytes()); + if (((bitField0_ & 0x00000010) != 0)) { + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(5, dagName_); } - if (((bitField0_ & 0x00000020) == 0x00000020)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(6, getVertexNameBytes()); + if (((bitField0_ & 0x00000020) != 0)) { + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(6, vertexName_); } - if (((bitField0_ & 0x00000040) == 0x00000040)) { + if (((bitField0_ & 0x00000040) != 0)) { size += com.google.protobuf.CodedOutputStream .computeInt32Size(7, vertexIndex_); } - if (((bitField0_ & 0x00000080) == 0x00000080)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(8, getTokenIdentifierBytes()); + if (((bitField0_ & 0x00000080) != 0)) { + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(8, tokenIdentifier_); } - if (((bitField0_ & 0x00000100) == 0x00000100)) { + if (((bitField0_ & 0x00000100) != 0)) { size += com.google.protobuf.CodedOutputStream - .computeMessageSize(9, processorDescriptor_); + .computeMessageSize(9, getProcessorDescriptor()); } for (int i = 0; i < inputSpecs_.size(); i++) { size += com.google.protobuf.CodedOutputStream @@ -4254,26 +4799,19 @@ public int getSerializedSize() { size += com.google.protobuf.CodedOutputStream .computeMessageSize(12, groupedInputSpecs_.get(i)); } - if (((bitField0_ & 0x00000200) == 0x00000200)) { + if (((bitField0_ & 0x00000200) != 0)) { size += com.google.protobuf.CodedOutputStream .computeInt32Size(13, vertexParallelism_); } - if (((bitField0_ & 0x00000400) == 0x00000400)) { + if (((bitField0_ & 0x00000400) != 0)) { size += com.google.protobuf.CodedOutputStream .computeBoolSize(14, isExternalSubmission_); } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { @@ -4284,88 +4822,86 @@ public boolean equals(final java.lang.Object obj) { } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec) obj; - boolean result = true; - result = result && (hasUser() == other.hasUser()); + if (hasUser() != other.hasUser()) return false; if (hasUser()) { - result = result && getUser() - .equals(other.getUser()); + if (!getUser() + .equals(other.getUser())) return false; } - result = result && (hasSignatureKeyId() == other.hasSignatureKeyId()); + if (hasSignatureKeyId() != other.hasSignatureKeyId()) return false; if (hasSignatureKeyId()) { - result = result && (getSignatureKeyId() - == other.getSignatureKeyId()); + if (getSignatureKeyId() + != other.getSignatureKeyId()) return false; } - result = result && (hasQueryIdentifier() == other.hasQueryIdentifier()); + if (hasQueryIdentifier() != other.hasQueryIdentifier()) return false; if (hasQueryIdentifier()) { - result = result && getQueryIdentifier() - .equals(other.getQueryIdentifier()); + if (!getQueryIdentifier() + .equals(other.getQueryIdentifier())) return false; } - result = result && (hasHiveQueryId() == other.hasHiveQueryId()); + if (hasHiveQueryId() != other.hasHiveQueryId()) return false; if (hasHiveQueryId()) { - result = result && getHiveQueryId() - .equals(other.getHiveQueryId()); + if (!getHiveQueryId() + .equals(other.getHiveQueryId())) return false; } - result = result && (hasDagName() == other.hasDagName()); + if (hasDagName() != other.hasDagName()) return false; if (hasDagName()) { - result = result && getDagName() - .equals(other.getDagName()); + if (!getDagName() + .equals(other.getDagName())) return false; } - result = result && (hasVertexName() == other.hasVertexName()); + if (hasVertexName() != other.hasVertexName()) return false; if (hasVertexName()) { - result = result && getVertexName() - .equals(other.getVertexName()); + if (!getVertexName() + .equals(other.getVertexName())) return false; } - result = result && (hasVertexIndex() == other.hasVertexIndex()); + if (hasVertexIndex() != other.hasVertexIndex()) return false; if (hasVertexIndex()) { - result = result && (getVertexIndex() - == other.getVertexIndex()); + if (getVertexIndex() + != other.getVertexIndex()) return false; } - result = result && (hasTokenIdentifier() == other.hasTokenIdentifier()); + if (hasTokenIdentifier() != other.hasTokenIdentifier()) return false; if (hasTokenIdentifier()) { - result = result && getTokenIdentifier() - .equals(other.getTokenIdentifier()); + if (!getTokenIdentifier() + .equals(other.getTokenIdentifier())) return false; } - result = result && (hasProcessorDescriptor() == other.hasProcessorDescriptor()); + if (hasProcessorDescriptor() != other.hasProcessorDescriptor()) return false; if (hasProcessorDescriptor()) { - result = result && getProcessorDescriptor() - .equals(other.getProcessorDescriptor()); - } - result = result && getInputSpecsList() - .equals(other.getInputSpecsList()); - result = result && getOutputSpecsList() - .equals(other.getOutputSpecsList()); - result = result && getGroupedInputSpecsList() - .equals(other.getGroupedInputSpecsList()); - result = result && (hasVertexParallelism() == other.hasVertexParallelism()); + if (!getProcessorDescriptor() + .equals(other.getProcessorDescriptor())) return false; + } + if (!getInputSpecsList() + .equals(other.getInputSpecsList())) return false; + if (!getOutputSpecsList() + .equals(other.getOutputSpecsList())) return false; + if (!getGroupedInputSpecsList() + .equals(other.getGroupedInputSpecsList())) return false; + if (hasVertexParallelism() != other.hasVertexParallelism()) return false; if (hasVertexParallelism()) { - result = result && (getVertexParallelism() - == other.getVertexParallelism()); + if (getVertexParallelism() + != other.getVertexParallelism()) return false; } - result = result && (hasIsExternalSubmission() == other.hasIsExternalSubmission()); + if (hasIsExternalSubmission() != other.hasIsExternalSubmission()) return false; if (hasIsExternalSubmission()) { - result = result && (getIsExternalSubmission() - == other.getIsExternalSubmission()); + if (getIsExternalSubmission() + != other.getIsExternalSubmission()) return false; } - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; + if (!unknownFields.equals(other.unknownFields)) return false; + return true; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); + hash = (19 * hash) + getDescriptor().hashCode(); if (hasUser()) { hash = (37 * hash) + USER_FIELD_NUMBER; hash = (53 * hash) + getUser().hashCode(); } if (hasSignatureKeyId()) { hash = (37 * hash) + SIGNATUREKEYID_FIELD_NUMBER; - hash = (53 * hash) + hashLong(getSignatureKeyId()); + hash = (53 * hash) + com.google.protobuf.Internal.hashLong( + getSignatureKeyId()); } if (hasQueryIdentifier()) { hash = (37 * hash) + QUERY_IDENTIFIER_FIELD_NUMBER; @@ -4413,13 +4949,25 @@ public int hashCode() { } if (hasIsExternalSubmission()) { hash = (37 * hash) + IS_EXTERNAL_SUBMISSION_FIELD_NUMBER; - hash = (53 * hash) + hashBoolean(getIsExternalSubmission()); + hash = (53 * hash) + com.google.protobuf.Internal.hashBoolean( + getIsExternalSubmission()); } - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec parseFrom( + java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec parseFrom( + java.nio.ByteBuffer data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { @@ -4443,65 +4991,80 @@ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.Si } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } + @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** - * Protobuf type {@code SignableVertexSpec} - * *
      * The part of SubmitWork that can be signed 
      * 
+ * + * Protobuf type {@code SignableVertexSpec} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpecOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder implements + // @@protoc_insertion_point(builder_implements:SignableVertexSpec) + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpecOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SignableVertexSpec_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SignableVertexSpec_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -4514,12 +5077,13 @@ private Builder() { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { getQueryIdentifierFieldBuilder(); getProcessorDescriptorFieldBuilder(); getInputSpecsFieldBuilder(); @@ -4527,10 +5091,7 @@ private void maybeForceBuilderInitialization() { getGroupedInputSpecsFieldBuilder(); } } - private static Builder create() { - return new Builder(); - } - + @java.lang.Override public Builder clear() { super.clear(); user_ = ""; @@ -4538,7 +5099,7 @@ public Builder clear() { signatureKeyId_ = 0L; bitField0_ = (bitField0_ & ~0x00000002); if (queryIdentifierBuilder_ == null) { - queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance(); + queryIdentifier_ = null; } else { queryIdentifierBuilder_.clear(); } @@ -4554,7 +5115,7 @@ public Builder clear() { tokenIdentifier_ = ""; bitField0_ = (bitField0_ & ~0x00000080); if (processorDescriptorBuilder_ == null) { - processorDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance(); + processorDescriptor_ = null; } else { processorDescriptorBuilder_.clear(); } @@ -4584,19 +5145,18 @@ public Builder clear() { return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - + @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SignableVertexSpec_descriptor; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.getDefaultInstance(); } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec result = buildPartial(); if (!result.isInitialized()) { @@ -4605,56 +5165,57 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableV return result; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { + if (((from_bitField0_ & 0x00000001) != 0)) { to_bitField0_ |= 0x00000001; } result.user_ = user_; - if (((from_bitField0_ & 0x00000002) == 0x00000002)) { + if (((from_bitField0_ & 0x00000002) != 0)) { + result.signatureKeyId_ = signatureKeyId_; to_bitField0_ |= 0x00000002; } - result.signatureKeyId_ = signatureKeyId_; - if (((from_bitField0_ & 0x00000004) == 0x00000004)) { + if (((from_bitField0_ & 0x00000004) != 0)) { + if (queryIdentifierBuilder_ == null) { + result.queryIdentifier_ = queryIdentifier_; + } else { + result.queryIdentifier_ = queryIdentifierBuilder_.build(); + } to_bitField0_ |= 0x00000004; } - if (queryIdentifierBuilder_ == null) { - result.queryIdentifier_ = queryIdentifier_; - } else { - result.queryIdentifier_ = queryIdentifierBuilder_.build(); - } - if (((from_bitField0_ & 0x00000008) == 0x00000008)) { + if (((from_bitField0_ & 0x00000008) != 0)) { to_bitField0_ |= 0x00000008; } result.hiveQueryId_ = hiveQueryId_; - if (((from_bitField0_ & 0x00000010) == 0x00000010)) { + if (((from_bitField0_ & 0x00000010) != 0)) { to_bitField0_ |= 0x00000010; } result.dagName_ = dagName_; - if (((from_bitField0_ & 0x00000020) == 0x00000020)) { + if (((from_bitField0_ & 0x00000020) != 0)) { to_bitField0_ |= 0x00000020; } result.vertexName_ = vertexName_; - if (((from_bitField0_ & 0x00000040) == 0x00000040)) { + if (((from_bitField0_ & 0x00000040) != 0)) { + result.vertexIndex_ = vertexIndex_; to_bitField0_ |= 0x00000040; } - result.vertexIndex_ = vertexIndex_; - if (((from_bitField0_ & 0x00000080) == 0x00000080)) { + if (((from_bitField0_ & 0x00000080) != 0)) { to_bitField0_ |= 0x00000080; } result.tokenIdentifier_ = tokenIdentifier_; - if (((from_bitField0_ & 0x00000100) == 0x00000100)) { + if (((from_bitField0_ & 0x00000100) != 0)) { + if (processorDescriptorBuilder_ == null) { + result.processorDescriptor_ = processorDescriptor_; + } else { + result.processorDescriptor_ = processorDescriptorBuilder_.build(); + } to_bitField0_ |= 0x00000100; } - if (processorDescriptorBuilder_ == null) { - result.processorDescriptor_ = processorDescriptor_; - } else { - result.processorDescriptor_ = processorDescriptorBuilder_.build(); - } if (inputSpecsBuilder_ == null) { - if (((bitField0_ & 0x00000200) == 0x00000200)) { + if (((bitField0_ & 0x00000200) != 0)) { inputSpecs_ = java.util.Collections.unmodifiableList(inputSpecs_); bitField0_ = (bitField0_ & ~0x00000200); } @@ -4663,7 +5224,7 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableV result.inputSpecs_ = inputSpecsBuilder_.build(); } if (outputSpecsBuilder_ == null) { - if (((bitField0_ & 0x00000400) == 0x00000400)) { + if (((bitField0_ & 0x00000400) != 0)) { outputSpecs_ = java.util.Collections.unmodifiableList(outputSpecs_); bitField0_ = (bitField0_ & ~0x00000400); } @@ -4672,7 +5233,7 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableV result.outputSpecs_ = outputSpecsBuilder_.build(); } if (groupedInputSpecsBuilder_ == null) { - if (((bitField0_ & 0x00000800) == 0x00000800)) { + if (((bitField0_ & 0x00000800) != 0)) { groupedInputSpecs_ = java.util.Collections.unmodifiableList(groupedInputSpecs_); bitField0_ = (bitField0_ & ~0x00000800); } @@ -4680,19 +5241,52 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableV } else { result.groupedInputSpecs_ = groupedInputSpecsBuilder_.build(); } - if (((from_bitField0_ & 0x00001000) == 0x00001000)) { + if (((from_bitField0_ & 0x00001000) != 0)) { + result.vertexParallelism_ = vertexParallelism_; to_bitField0_ |= 0x00000200; } - result.vertexParallelism_ = vertexParallelism_; - if (((from_bitField0_ & 0x00002000) == 0x00002000)) { + if (((from_bitField0_ & 0x00002000) != 0)) { + result.isExternalSubmission_ = isExternalSubmission_; to_bitField0_ |= 0x00000400; } - result.isExternalSubmission_ = isExternalSubmission_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } + @java.lang.Override + public Builder clone() { + return super.clone(); + } + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.setField(field, value); + } + @java.lang.Override + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + @java.lang.Override + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.addRepeatedField(field, value); + } + @java.lang.Override public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec)other); @@ -4760,7 +5354,7 @@ public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtoc inputSpecs_ = other.inputSpecs_; bitField0_ = (bitField0_ & ~0x00000200); inputSpecsBuilder_ = - com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? + com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getInputSpecsFieldBuilder() : null; } else { inputSpecsBuilder_.addAllMessages(other.inputSpecs_); @@ -4786,7 +5380,7 @@ public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtoc outputSpecs_ = other.outputSpecs_; bitField0_ = (bitField0_ & ~0x00000400); outputSpecsBuilder_ = - com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? + com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getOutputSpecsFieldBuilder() : null; } else { outputSpecsBuilder_.addAllMessages(other.outputSpecs_); @@ -4812,7 +5406,7 @@ public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtoc groupedInputSpecs_ = other.groupedInputSpecs_; bitField0_ = (bitField0_ & ~0x00000800); groupedInputSpecsBuilder_ = - com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? + com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getGroupedInputSpecsFieldBuilder() : null; } else { groupedInputSpecsBuilder_.addAllMessages(other.groupedInputSpecs_); @@ -4825,14 +5419,17 @@ public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtoc if (other.hasIsExternalSubmission()) { setIsExternalSubmission(other.getIsExternalSubmission()); } - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } + @java.lang.Override public final boolean isInitialized() { return true; } + @java.lang.Override public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) @@ -4842,7 +5439,7 @@ public Builder mergeFrom( parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -4852,23 +5449,27 @@ public Builder mergeFrom( } private int bitField0_; - // optional string user = 1; private java.lang.Object user_ = ""; /** * optional string user = 1; + * @return Whether the user field is set. */ public boolean hasUser() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional string user = 1; + * @return The user. */ public java.lang.String getUser() { java.lang.Object ref = user_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - user_ = s; + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + user_ = s; + } return s; } else { return (java.lang.String) ref; @@ -4876,6 +5477,7 @@ public java.lang.String getUser() { } /** * optional string user = 1; + * @return The bytes for user. */ public com.google.protobuf.ByteString getUserBytes() { @@ -4892,6 +5494,8 @@ public java.lang.String getUser() { } /** * optional string user = 1; + * @param value The user to set. + * @return This builder for chaining. */ public Builder setUser( java.lang.String value) { @@ -4905,6 +5509,7 @@ public Builder setUser( } /** * optional string user = 1; + * @return This builder for chaining. */ public Builder clearUser() { bitField0_ = (bitField0_ & ~0x00000001); @@ -4914,6 +5519,8 @@ public Builder clearUser() { } /** * optional string user = 1; + * @param value The bytes for user to set. + * @return This builder for chaining. */ public Builder setUserBytes( com.google.protobuf.ByteString value) { @@ -4926,22 +5533,27 @@ public Builder setUserBytes( return this; } - // optional int64 signatureKeyId = 2; private long signatureKeyId_ ; /** * optional int64 signatureKeyId = 2; + * @return Whether the signatureKeyId field is set. */ + @java.lang.Override public boolean hasSignatureKeyId() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * optional int64 signatureKeyId = 2; + * @return The signatureKeyId. */ + @java.lang.Override public long getSignatureKeyId() { return signatureKeyId_; } /** * optional int64 signatureKeyId = 2; + * @param value The signatureKeyId to set. + * @return This builder for chaining. */ public Builder setSignatureKeyId(long value) { bitField0_ |= 0x00000002; @@ -4951,6 +5563,7 @@ public Builder setSignatureKeyId(long value) { } /** * optional int64 signatureKeyId = 2; + * @return This builder for chaining. */ public Builder clearSignatureKeyId() { bitField0_ = (bitField0_ & ~0x00000002); @@ -4959,22 +5572,23 @@ public Builder clearSignatureKeyId() { return this; } - // optional .QueryIdentifierProto query_identifier = 3; - private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance(); - private com.google.protobuf.SingleFieldBuilder< + private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto queryIdentifier_; + private com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder> queryIdentifierBuilder_; /** * optional .QueryIdentifierProto query_identifier = 3; + * @return Whether the queryIdentifier field is set. */ public boolean hasQueryIdentifier() { - return ((bitField0_ & 0x00000004) == 0x00000004); + return ((bitField0_ & 0x00000004) != 0); } /** * optional .QueryIdentifierProto query_identifier = 3; + * @return The queryIdentifier. */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto getQueryIdentifier() { if (queryIdentifierBuilder_ == null) { - return queryIdentifier_; + return queryIdentifier_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance() : queryIdentifier_; } else { return queryIdentifierBuilder_.getMessage(); } @@ -5014,7 +5628,8 @@ public Builder setQueryIdentifier( */ public Builder mergeQueryIdentifier(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto value) { if (queryIdentifierBuilder_ == null) { - if (((bitField0_ & 0x00000004) == 0x00000004) && + if (((bitField0_ & 0x00000004) != 0) && + queryIdentifier_ != null && queryIdentifier_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance()) { queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.newBuilder(queryIdentifier_).mergeFrom(value).buildPartial(); @@ -5033,7 +5648,7 @@ public Builder mergeQueryIdentifier(org.apache.hadoop.hive.llap.daemon.rpc.LlapD */ public Builder clearQueryIdentifier() { if (queryIdentifierBuilder_ == null) { - queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance(); + queryIdentifier_ = null; onChanged(); } else { queryIdentifierBuilder_.clear(); @@ -5056,19 +5671,20 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIden if (queryIdentifierBuilder_ != null) { return queryIdentifierBuilder_.getMessageOrBuilder(); } else { - return queryIdentifier_; + return queryIdentifier_ == null ? + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance() : queryIdentifier_; } } /** * optional .QueryIdentifierProto query_identifier = 3; */ - private com.google.protobuf.SingleFieldBuilder< + private com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder> getQueryIdentifierFieldBuilder() { if (queryIdentifierBuilder_ == null) { - queryIdentifierBuilder_ = new com.google.protobuf.SingleFieldBuilder< + queryIdentifierBuilder_ = new com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder>( - queryIdentifier_, + getQueryIdentifier(), getParentForChildren(), isClean()); queryIdentifier_ = null; @@ -5076,23 +5692,27 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIden return queryIdentifierBuilder_; } - // optional string hive_query_id = 4; private java.lang.Object hiveQueryId_ = ""; /** * optional string hive_query_id = 4; + * @return Whether the hiveQueryId field is set. */ public boolean hasHiveQueryId() { - return ((bitField0_ & 0x00000008) == 0x00000008); + return ((bitField0_ & 0x00000008) != 0); } /** * optional string hive_query_id = 4; + * @return The hiveQueryId. */ public java.lang.String getHiveQueryId() { java.lang.Object ref = hiveQueryId_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - hiveQueryId_ = s; + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + hiveQueryId_ = s; + } return s; } else { return (java.lang.String) ref; @@ -5100,6 +5720,7 @@ public java.lang.String getHiveQueryId() { } /** * optional string hive_query_id = 4; + * @return The bytes for hiveQueryId. */ public com.google.protobuf.ByteString getHiveQueryIdBytes() { @@ -5116,6 +5737,8 @@ public java.lang.String getHiveQueryId() { } /** * optional string hive_query_id = 4; + * @param value The hiveQueryId to set. + * @return This builder for chaining. */ public Builder setHiveQueryId( java.lang.String value) { @@ -5129,6 +5752,7 @@ public Builder setHiveQueryId( } /** * optional string hive_query_id = 4; + * @return This builder for chaining. */ public Builder clearHiveQueryId() { bitField0_ = (bitField0_ & ~0x00000008); @@ -5138,6 +5762,8 @@ public Builder clearHiveQueryId() { } /** * optional string hive_query_id = 4; + * @param value The bytes for hiveQueryId to set. + * @return This builder for chaining. */ public Builder setHiveQueryIdBytes( com.google.protobuf.ByteString value) { @@ -5150,42 +5776,47 @@ public Builder setHiveQueryIdBytes( return this; } - // optional string dag_name = 5; private java.lang.Object dagName_ = ""; /** - * optional string dag_name = 5; - * *
        * Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here.
        * 
+ * + * optional string dag_name = 5; + * @return Whether the dagName field is set. */ public boolean hasDagName() { - return ((bitField0_ & 0x00000010) == 0x00000010); + return ((bitField0_ & 0x00000010) != 0); } /** - * optional string dag_name = 5; - * *
        * Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here.
        * 
+ * + * optional string dag_name = 5; + * @return The dagName. */ public java.lang.String getDagName() { java.lang.Object ref = dagName_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - dagName_ = s; + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + dagName_ = s; + } return s; } else { return (java.lang.String) ref; } } /** - * optional string dag_name = 5; - * *
        * Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here.
        * 
+ * + * optional string dag_name = 5; + * @return The bytes for dagName. */ public com.google.protobuf.ByteString getDagNameBytes() { @@ -5201,11 +5832,13 @@ public java.lang.String getDagName() { } } /** - * optional string dag_name = 5; - * *
        * Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here.
        * 
+ * + * optional string dag_name = 5; + * @param value The dagName to set. + * @return This builder for chaining. */ public Builder setDagName( java.lang.String value) { @@ -5218,11 +5851,12 @@ public Builder setDagName( return this; } /** - * optional string dag_name = 5; - * *
        * Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here.
        * 
+ * + * optional string dag_name = 5; + * @return This builder for chaining. */ public Builder clearDagName() { bitField0_ = (bitField0_ & ~0x00000010); @@ -5231,11 +5865,13 @@ public Builder clearDagName() { return this; } /** - * optional string dag_name = 5; - * *
        * Display names cannot be modified by the client for now. If needed, they should be sent to HS2 who will put them here.
        * 
+ * + * optional string dag_name = 5; + * @param value The bytes for dagName to set. + * @return This builder for chaining. */ public Builder setDagNameBytes( com.google.protobuf.ByteString value) { @@ -5248,23 +5884,27 @@ public Builder setDagNameBytes( return this; } - // optional string vertex_name = 6; private java.lang.Object vertexName_ = ""; /** * optional string vertex_name = 6; + * @return Whether the vertexName field is set. */ public boolean hasVertexName() { - return ((bitField0_ & 0x00000020) == 0x00000020); + return ((bitField0_ & 0x00000020) != 0); } /** * optional string vertex_name = 6; + * @return The vertexName. */ public java.lang.String getVertexName() { java.lang.Object ref = vertexName_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - vertexName_ = s; + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + vertexName_ = s; + } return s; } else { return (java.lang.String) ref; @@ -5272,6 +5912,7 @@ public java.lang.String getVertexName() { } /** * optional string vertex_name = 6; + * @return The bytes for vertexName. */ public com.google.protobuf.ByteString getVertexNameBytes() { @@ -5288,6 +5929,8 @@ public java.lang.String getVertexName() { } /** * optional string vertex_name = 6; + * @param value The vertexName to set. + * @return This builder for chaining. */ public Builder setVertexName( java.lang.String value) { @@ -5301,6 +5944,7 @@ public Builder setVertexName( } /** * optional string vertex_name = 6; + * @return This builder for chaining. */ public Builder clearVertexName() { bitField0_ = (bitField0_ & ~0x00000020); @@ -5310,6 +5954,8 @@ public Builder clearVertexName() { } /** * optional string vertex_name = 6; + * @param value The bytes for vertexName to set. + * @return This builder for chaining. */ public Builder setVertexNameBytes( com.google.protobuf.ByteString value) { @@ -5322,22 +5968,27 @@ public Builder setVertexNameBytes( return this; } - // optional int32 vertex_index = 7; private int vertexIndex_ ; /** * optional int32 vertex_index = 7; + * @return Whether the vertexIndex field is set. */ + @java.lang.Override public boolean hasVertexIndex() { - return ((bitField0_ & 0x00000040) == 0x00000040); + return ((bitField0_ & 0x00000040) != 0); } /** * optional int32 vertex_index = 7; + * @return The vertexIndex. */ + @java.lang.Override public int getVertexIndex() { return vertexIndex_; } /** * optional int32 vertex_index = 7; + * @param value The vertexIndex to set. + * @return This builder for chaining. */ public Builder setVertexIndex(int value) { bitField0_ |= 0x00000040; @@ -5347,6 +5998,7 @@ public Builder setVertexIndex(int value) { } /** * optional int32 vertex_index = 7; + * @return This builder for chaining. */ public Builder clearVertexIndex() { bitField0_ = (bitField0_ & ~0x00000040); @@ -5355,42 +6007,47 @@ public Builder clearVertexIndex() { return this; } - // optional string token_identifier = 8; private java.lang.Object tokenIdentifier_ = ""; /** - * optional string token_identifier = 8; - * *
        * The core vertex stuff 
        * 
+ * + * optional string token_identifier = 8; + * @return Whether the tokenIdentifier field is set. */ public boolean hasTokenIdentifier() { - return ((bitField0_ & 0x00000080) == 0x00000080); + return ((bitField0_ & 0x00000080) != 0); } /** - * optional string token_identifier = 8; - * *
        * The core vertex stuff 
        * 
+ * + * optional string token_identifier = 8; + * @return The tokenIdentifier. */ public java.lang.String getTokenIdentifier() { java.lang.Object ref = tokenIdentifier_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - tokenIdentifier_ = s; + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + tokenIdentifier_ = s; + } return s; } else { return (java.lang.String) ref; } } /** - * optional string token_identifier = 8; - * *
        * The core vertex stuff 
        * 
+ * + * optional string token_identifier = 8; + * @return The bytes for tokenIdentifier. */ public com.google.protobuf.ByteString getTokenIdentifierBytes() { @@ -5406,11 +6063,13 @@ public java.lang.String getTokenIdentifier() { } } /** - * optional string token_identifier = 8; - * *
        * The core vertex stuff 
        * 
+ * + * optional string token_identifier = 8; + * @param value The tokenIdentifier to set. + * @return This builder for chaining. */ public Builder setTokenIdentifier( java.lang.String value) { @@ -5423,11 +6082,12 @@ public Builder setTokenIdentifier( return this; } /** - * optional string token_identifier = 8; - * *
        * The core vertex stuff 
        * 
+ * + * optional string token_identifier = 8; + * @return This builder for chaining. */ public Builder clearTokenIdentifier() { bitField0_ = (bitField0_ & ~0x00000080); @@ -5436,11 +6096,13 @@ public Builder clearTokenIdentifier() { return this; } /** - * optional string token_identifier = 8; - * *
        * The core vertex stuff 
        * 
+ * + * optional string token_identifier = 8; + * @param value The bytes for tokenIdentifier to set. + * @return This builder for chaining. */ public Builder setTokenIdentifierBytes( com.google.protobuf.ByteString value) { @@ -5453,22 +6115,23 @@ public Builder setTokenIdentifierBytes( return this; } - // optional .EntityDescriptorProto processor_descriptor = 9; - private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto processorDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance(); - private com.google.protobuf.SingleFieldBuilder< + private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto processorDescriptor_; + private com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder> processorDescriptorBuilder_; /** * optional .EntityDescriptorProto processor_descriptor = 9; + * @return Whether the processorDescriptor field is set. */ public boolean hasProcessorDescriptor() { - return ((bitField0_ & 0x00000100) == 0x00000100); + return ((bitField0_ & 0x00000100) != 0); } /** * optional .EntityDescriptorProto processor_descriptor = 9; + * @return The processorDescriptor. */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto getProcessorDescriptor() { if (processorDescriptorBuilder_ == null) { - return processorDescriptor_; + return processorDescriptor_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance() : processorDescriptor_; } else { return processorDescriptorBuilder_.getMessage(); } @@ -5508,7 +6171,8 @@ public Builder setProcessorDescriptor( */ public Builder mergeProcessorDescriptor(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto value) { if (processorDescriptorBuilder_ == null) { - if (((bitField0_ & 0x00000100) == 0x00000100) && + if (((bitField0_ & 0x00000100) != 0) && + processorDescriptor_ != null && processorDescriptor_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance()) { processorDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.newBuilder(processorDescriptor_).mergeFrom(value).buildPartial(); @@ -5527,7 +6191,7 @@ public Builder mergeProcessorDescriptor(org.apache.hadoop.hive.llap.daemon.rpc.L */ public Builder clearProcessorDescriptor() { if (processorDescriptorBuilder_ == null) { - processorDescriptor_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance(); + processorDescriptor_ = null; onChanged(); } else { processorDescriptorBuilder_.clear(); @@ -5550,19 +6214,20 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDes if (processorDescriptorBuilder_ != null) { return processorDescriptorBuilder_.getMessageOrBuilder(); } else { - return processorDescriptor_; + return processorDescriptor_ == null ? + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.getDefaultInstance() : processorDescriptor_; } } /** * optional .EntityDescriptorProto processor_descriptor = 9; */ - private com.google.protobuf.SingleFieldBuilder< + private com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder> getProcessorDescriptorFieldBuilder() { if (processorDescriptorBuilder_ == null) { - processorDescriptorBuilder_ = new com.google.protobuf.SingleFieldBuilder< + processorDescriptorBuilder_ = new com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDescriptorProtoOrBuilder>( - processorDescriptor_, + getProcessorDescriptor(), getParentForChildren(), isClean()); processorDescriptor_ = null; @@ -5570,17 +6235,16 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EntityDes return processorDescriptorBuilder_; } - // repeated .IOSpecProto input_specs = 10; private java.util.List inputSpecs_ = java.util.Collections.emptyList(); private void ensureInputSpecsIsMutable() { - if (!((bitField0_ & 0x00000200) == 0x00000200)) { + if (!((bitField0_ & 0x00000200) != 0)) { inputSpecs_ = new java.util.ArrayList(inputSpecs_); bitField0_ |= 0x00000200; } } - private com.google.protobuf.RepeatedFieldBuilder< + private com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder> inputSpecsBuilder_; /** @@ -5712,7 +6376,8 @@ public Builder addAllInputSpecs( java.lang.Iterable values) { if (inputSpecsBuilder_ == null) { ensureInputSpecsIsMutable(); - super.addAll(values, inputSpecs_); + com.google.protobuf.AbstractMessageLite.Builder.addAll( + values, inputSpecs_); onChanged(); } else { inputSpecsBuilder_.addAllMessages(values); @@ -5795,14 +6460,14 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecPro getInputSpecsBuilderList() { return getInputSpecsFieldBuilder().getBuilderList(); } - private com.google.protobuf.RepeatedFieldBuilder< + private com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder> getInputSpecsFieldBuilder() { if (inputSpecsBuilder_ == null) { - inputSpecsBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< + inputSpecsBuilder_ = new com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder>( inputSpecs_, - ((bitField0_ & 0x00000200) == 0x00000200), + ((bitField0_ & 0x00000200) != 0), getParentForChildren(), isClean()); inputSpecs_ = null; @@ -5810,17 +6475,16 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecPro return inputSpecsBuilder_; } - // repeated .IOSpecProto output_specs = 11; private java.util.List outputSpecs_ = java.util.Collections.emptyList(); private void ensureOutputSpecsIsMutable() { - if (!((bitField0_ & 0x00000400) == 0x00000400)) { + if (!((bitField0_ & 0x00000400) != 0)) { outputSpecs_ = new java.util.ArrayList(outputSpecs_); bitField0_ |= 0x00000400; } } - private com.google.protobuf.RepeatedFieldBuilder< + private com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder> outputSpecsBuilder_; /** @@ -5952,7 +6616,8 @@ public Builder addAllOutputSpecs( java.lang.Iterable values) { if (outputSpecsBuilder_ == null) { ensureOutputSpecsIsMutable(); - super.addAll(values, outputSpecs_); + com.google.protobuf.AbstractMessageLite.Builder.addAll( + values, outputSpecs_); onChanged(); } else { outputSpecsBuilder_.addAllMessages(values); @@ -6035,14 +6700,14 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecPro getOutputSpecsBuilderList() { return getOutputSpecsFieldBuilder().getBuilderList(); } - private com.google.protobuf.RepeatedFieldBuilder< + private com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder> getOutputSpecsFieldBuilder() { if (outputSpecsBuilder_ == null) { - outputSpecsBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< + outputSpecsBuilder_ = new com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecProtoOrBuilder>( outputSpecs_, - ((bitField0_ & 0x00000400) == 0x00000400), + ((bitField0_ & 0x00000400) != 0), getParentForChildren(), isClean()); outputSpecs_ = null; @@ -6050,17 +6715,16 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.IOSpecPro return outputSpecsBuilder_; } - // repeated .GroupInputSpecProto grouped_input_specs = 12; private java.util.List groupedInputSpecs_ = java.util.Collections.emptyList(); private void ensureGroupedInputSpecsIsMutable() { - if (!((bitField0_ & 0x00000800) == 0x00000800)) { + if (!((bitField0_ & 0x00000800) != 0)) { groupedInputSpecs_ = new java.util.ArrayList(groupedInputSpecs_); bitField0_ |= 0x00000800; } } - private com.google.protobuf.RepeatedFieldBuilder< + private com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProtoOrBuilder> groupedInputSpecsBuilder_; /** @@ -6192,7 +6856,8 @@ public Builder addAllGroupedInputSpecs( java.lang.Iterable values) { if (groupedInputSpecsBuilder_ == null) { ensureGroupedInputSpecsIsMutable(); - super.addAll(values, groupedInputSpecs_); + com.google.protobuf.AbstractMessageLite.Builder.addAll( + values, groupedInputSpecs_); onChanged(); } else { groupedInputSpecsBuilder_.addAllMessages(values); @@ -6275,14 +6940,14 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInpu getGroupedInputSpecsBuilderList() { return getGroupedInputSpecsFieldBuilder().getBuilderList(); } - private com.google.protobuf.RepeatedFieldBuilder< + private com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProtoOrBuilder> getGroupedInputSpecsFieldBuilder() { if (groupedInputSpecsBuilder_ == null) { - groupedInputSpecsBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< + groupedInputSpecsBuilder_ = new com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInputSpecProtoOrBuilder>( groupedInputSpecs_, - ((bitField0_ & 0x00000800) == 0x00000800), + ((bitField0_ & 0x00000800) != 0), getParentForChildren(), isClean()); groupedInputSpecs_ = null; @@ -6290,34 +6955,39 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GroupInpu return groupedInputSpecsBuilder_; } - // optional int32 vertex_parallelism = 13; private int vertexParallelism_ ; /** - * optional int32 vertex_parallelism = 13; - * *
        * An internal field required for Tez.
        * 
+ * + * optional int32 vertex_parallelism = 13; + * @return Whether the vertexParallelism field is set. */ + @java.lang.Override public boolean hasVertexParallelism() { - return ((bitField0_ & 0x00001000) == 0x00001000); + return ((bitField0_ & 0x00001000) != 0); } /** - * optional int32 vertex_parallelism = 13; - * *
        * An internal field required for Tez.
        * 
+ * + * optional int32 vertex_parallelism = 13; + * @return The vertexParallelism. */ + @java.lang.Override public int getVertexParallelism() { return vertexParallelism_; } /** - * optional int32 vertex_parallelism = 13; - * *
        * An internal field required for Tez.
        * 
+ * + * optional int32 vertex_parallelism = 13; + * @param value The vertexParallelism to set. + * @return This builder for chaining. */ public Builder setVertexParallelism(int value) { bitField0_ |= 0x00001000; @@ -6326,11 +6996,12 @@ public Builder setVertexParallelism(int value) { return this; } /** - * optional int32 vertex_parallelism = 13; - * *
        * An internal field required for Tez.
        * 
+ * + * optional int32 vertex_parallelism = 13; + * @return This builder for chaining. */ public Builder clearVertexParallelism() { bitField0_ = (bitField0_ & ~0x00001000); @@ -6339,22 +7010,27 @@ public Builder clearVertexParallelism() { return this; } - // optional bool is_external_submission = 14 [default = false]; private boolean isExternalSubmission_ ; /** * optional bool is_external_submission = 14 [default = false]; + * @return Whether the isExternalSubmission field is set. */ + @java.lang.Override public boolean hasIsExternalSubmission() { - return ((bitField0_ & 0x00002000) == 0x00002000); + return ((bitField0_ & 0x00002000) != 0); } /** * optional bool is_external_submission = 14 [default = false]; + * @return The isExternalSubmission. */ + @java.lang.Override public boolean getIsExternalSubmission() { return isExternalSubmission_; } /** * optional bool is_external_submission = 14 [default = false]; + * @param value The isExternalSubmission to set. + * @return This builder for chaining. */ public Builder setIsExternalSubmission(boolean value) { bitField0_ |= 0x00002000; @@ -6364,6 +7040,7 @@ public Builder setIsExternalSubmission(boolean value) { } /** * optional bool is_external_submission = 14 [default = false]; + * @return This builder for chaining. */ public Builder clearIsExternalSubmission() { bitField0_ = (bitField0_ & ~0x00002000); @@ -6371,28 +7048,71 @@ public Builder clearIsExternalSubmission() { onChanged(); return this; } + @java.lang.Override + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:SignableVertexSpec) } + // @@protoc_insertion_point(class_scope:SignableVertexSpec) + private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec DEFAULT_INSTANCE; static { - defaultInstance = new SignableVertexSpec(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec(); + } + + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser + PARSER = new com.google.protobuf.AbstractParser() { + @java.lang.Override + public SignableVertexSpec parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new SignableVertexSpec(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec getDefaultInstanceForType() { + return DEFAULT_INSTANCE; } - // @@protoc_insertion_point(class_scope:SignableVertexSpec) } - public interface VertexOrBinaryOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface VertexOrBinaryOrBuilder extends + // @@protoc_insertion_point(interface_extends:VertexOrBinary) + com.google.protobuf.MessageOrBuilder { - // optional .SignableVertexSpec vertex = 1; /** * optional .SignableVertexSpec vertex = 1; + * @return Whether the vertex field is set. */ boolean hasVertex(); /** * optional .SignableVertexSpec vertex = 1; + * @return The vertex. */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec getVertex(); /** @@ -6400,61 +7120,65 @@ public interface VertexOrBinaryOrBuilder */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpecOrBuilder getVertexOrBuilder(); - // optional bytes vertexBinary = 2; /** - * optional bytes vertexBinary = 2; - * *
      * SignableVertexSpec
      * 
+ * + * optional bytes vertexBinary = 2; + * @return Whether the vertexBinary field is set. */ boolean hasVertexBinary(); /** - * optional bytes vertexBinary = 2; - * *
      * SignableVertexSpec
      * 
+ * + * optional bytes vertexBinary = 2; + * @return The vertexBinary. */ com.google.protobuf.ByteString getVertexBinary(); } /** - * Protobuf type {@code VertexOrBinary} - * *
    * Union
    * 
+ * + * Protobuf type {@code VertexOrBinary} */ public static final class VertexOrBinary extends - com.google.protobuf.GeneratedMessage - implements VertexOrBinaryOrBuilder { + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:VertexOrBinary) + VertexOrBinaryOrBuilder { + private static final long serialVersionUID = 0L; // Use VertexOrBinary.newBuilder() to construct. - private VertexOrBinary(com.google.protobuf.GeneratedMessage.Builder builder) { + private VertexOrBinary(com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private VertexOrBinary(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final VertexOrBinary defaultInstance; - public static VertexOrBinary getDefaultInstance() { - return defaultInstance; + private VertexOrBinary() { + vertexBinary_ = com.google.protobuf.ByteString.EMPTY; } - public VertexOrBinary getDefaultInstanceForType() { - return defaultInstance; + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance( + UnusedPrivateParameter unused) { + return new VertexOrBinary(); } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private VertexOrBinary( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); @@ -6466,16 +7190,9 @@ private VertexOrBinary( case 0: done = true; break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } case 10: { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.Builder subBuilder = null; - if (((bitField0_ & 0x00000001) == 0x00000001)) { + if (((bitField0_ & 0x00000001) != 0)) { subBuilder = vertex_.toBuilder(); } vertex_ = input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.PARSER, extensionRegistry); @@ -6491,13 +7208,20 @@ private VertexOrBinary( vertexBinary_ = input.readBytes(); break; } + default: { + if (!parseUnknownField( + input, unknownFields, extensionRegistry, tag)) { + done = true; + } + break; + } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); @@ -6508,126 +7232,110 @@ private VertexOrBinary( return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_VertexOrBinary_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_VertexOrBinary_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public VertexOrBinary parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new VertexOrBinary(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - private int bitField0_; - // optional .SignableVertexSpec vertex = 1; public static final int VERTEX_FIELD_NUMBER = 1; private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec vertex_; /** * optional .SignableVertexSpec vertex = 1; + * @return Whether the vertex field is set. */ + @java.lang.Override public boolean hasVertex() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional .SignableVertexSpec vertex = 1; + * @return The vertex. */ + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec getVertex() { - return vertex_; + return vertex_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.getDefaultInstance() : vertex_; } /** * optional .SignableVertexSpec vertex = 1; */ + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpecOrBuilder getVertexOrBuilder() { - return vertex_; + return vertex_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.getDefaultInstance() : vertex_; } - // optional bytes vertexBinary = 2; public static final int VERTEXBINARY_FIELD_NUMBER = 2; private com.google.protobuf.ByteString vertexBinary_; /** - * optional bytes vertexBinary = 2; - * *
      * SignableVertexSpec
      * 
- */ + * + * optional bytes vertexBinary = 2; + * @return Whether the vertexBinary field is set. + */ + @java.lang.Override public boolean hasVertexBinary() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** - * optional bytes vertexBinary = 2; - * *
      * SignableVertexSpec
      * 
+ * + * optional bytes vertexBinary = 2; + * @return The vertexBinary. */ + @java.lang.Override public com.google.protobuf.ByteString getVertexBinary() { return vertexBinary_; } - private void initFields() { - vertex_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.getDefaultInstance(); - vertexBinary_ = com.google.protobuf.ByteString.EMPTY; - } private byte memoizedIsInitialized = -1; + @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } + @java.lang.Override public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); - if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeMessage(1, vertex_); + if (((bitField0_ & 0x00000001) != 0)) { + output.writeMessage(1, getVertex()); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { + if (((bitField0_ & 0x00000002) != 0)) { output.writeBytes(2, vertexBinary_); } - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; + @java.lang.Override public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { + if (((bitField0_ & 0x00000001) != 0)) { size += com.google.protobuf.CodedOutputStream - .computeMessageSize(1, vertex_); + .computeMessageSize(1, getVertex()); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { + if (((bitField0_ & 0x00000002) != 0)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(2, vertexBinary_); } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { @@ -6638,30 +7346,27 @@ public boolean equals(final java.lang.Object obj) { } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary) obj; - boolean result = true; - result = result && (hasVertex() == other.hasVertex()); + if (hasVertex() != other.hasVertex()) return false; if (hasVertex()) { - result = result && getVertex() - .equals(other.getVertex()); + if (!getVertex() + .equals(other.getVertex())) return false; } - result = result && (hasVertexBinary() == other.hasVertexBinary()); + if (hasVertexBinary() != other.hasVertexBinary()) return false; if (hasVertexBinary()) { - result = result && getVertexBinary() - .equals(other.getVertexBinary()); + if (!getVertexBinary() + .equals(other.getVertexBinary())) return false; } - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; + if (!unknownFields.equals(other.unknownFields)) return false; + return true; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); + hash = (19 * hash) + getDescriptor().hashCode(); if (hasVertex()) { hash = (37 * hash) + VERTEX_FIELD_NUMBER; hash = (53 * hash) + getVertex().hashCode(); @@ -6670,11 +7375,22 @@ public int hashCode() { hash = (37 * hash) + VERTEXBINARY_FIELD_NUMBER; hash = (53 * hash) + getVertexBinary().hashCode(); } - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary parseFrom( + java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary parseFrom( + java.nio.ByteBuffer data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { @@ -6698,65 +7414,80 @@ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.Ve } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } + @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** - * Protobuf type {@code VertexOrBinary} - * *
      * Union
      * 
+ * + * Protobuf type {@code VertexOrBinary} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinaryOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder implements + // @@protoc_insertion_point(builder_implements:VertexOrBinary) + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinaryOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_VertexOrBinary_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_VertexOrBinary_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -6769,23 +7500,21 @@ private Builder() { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { getVertexFieldBuilder(); } } - private static Builder create() { - return new Builder(); - } - + @java.lang.Override public Builder clear() { super.clear(); if (vertexBuilder_ == null) { - vertex_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.getDefaultInstance(); + vertex_ = null; } else { vertexBuilder_.clear(); } @@ -6795,19 +7524,18 @@ public Builder clear() { return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - + @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_VertexOrBinary_descriptor; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary.getDefaultInstance(); } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary result = buildPartial(); if (!result.isInitialized()) { @@ -6816,19 +7544,20 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrB return result; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { + if (((from_bitField0_ & 0x00000001) != 0)) { + if (vertexBuilder_ == null) { + result.vertex_ = vertex_; + } else { + result.vertex_ = vertexBuilder_.build(); + } to_bitField0_ |= 0x00000001; } - if (vertexBuilder_ == null) { - result.vertex_ = vertex_; - } else { - result.vertex_ = vertexBuilder_.build(); - } - if (((from_bitField0_ & 0x00000002) == 0x00000002)) { + if (((from_bitField0_ & 0x00000002) != 0)) { to_bitField0_ |= 0x00000002; } result.vertexBinary_ = vertexBinary_; @@ -6837,6 +7566,39 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrB return result; } + @java.lang.Override + public Builder clone() { + return super.clone(); + } + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.setField(field, value); + } + @java.lang.Override + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + @java.lang.Override + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.addRepeatedField(field, value); + } + @java.lang.Override public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary)other); @@ -6854,14 +7616,17 @@ public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtoc if (other.hasVertexBinary()) { setVertexBinary(other.getVertexBinary()); } - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } + @java.lang.Override public final boolean isInitialized() { return true; } + @java.lang.Override public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) @@ -6871,7 +7636,7 @@ public Builder mergeFrom( parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -6881,22 +7646,23 @@ public Builder mergeFrom( } private int bitField0_; - // optional .SignableVertexSpec vertex = 1; - private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec vertex_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.getDefaultInstance(); - private com.google.protobuf.SingleFieldBuilder< + private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec vertex_; + private com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpecOrBuilder> vertexBuilder_; /** * optional .SignableVertexSpec vertex = 1; + * @return Whether the vertex field is set. */ public boolean hasVertex() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional .SignableVertexSpec vertex = 1; + * @return The vertex. */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec getVertex() { if (vertexBuilder_ == null) { - return vertex_; + return vertex_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.getDefaultInstance() : vertex_; } else { return vertexBuilder_.getMessage(); } @@ -6936,7 +7702,8 @@ public Builder setVertex( */ public Builder mergeVertex(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec value) { if (vertexBuilder_ == null) { - if (((bitField0_ & 0x00000001) == 0x00000001) && + if (((bitField0_ & 0x00000001) != 0) && + vertex_ != null && vertex_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.getDefaultInstance()) { vertex_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.newBuilder(vertex_).mergeFrom(value).buildPartial(); @@ -6955,7 +7722,7 @@ public Builder mergeVertex(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProt */ public Builder clearVertex() { if (vertexBuilder_ == null) { - vertex_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.getDefaultInstance(); + vertex_ = null; onChanged(); } else { vertexBuilder_.clear(); @@ -6978,19 +7745,20 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableV if (vertexBuilder_ != null) { return vertexBuilder_.getMessageOrBuilder(); } else { - return vertex_; + return vertex_ == null ? + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.getDefaultInstance() : vertex_; } } /** * optional .SignableVertexSpec vertex = 1; */ - private com.google.protobuf.SingleFieldBuilder< + private com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpecOrBuilder> getVertexFieldBuilder() { if (vertexBuilder_ == null) { - vertexBuilder_ = new com.google.protobuf.SingleFieldBuilder< + vertexBuilder_ = new com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpec.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableVertexSpecOrBuilder>( - vertex_, + getVertex(), getParentForChildren(), isClean()); vertex_ = null; @@ -6998,34 +7766,39 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SignableV return vertexBuilder_; } - // optional bytes vertexBinary = 2; private com.google.protobuf.ByteString vertexBinary_ = com.google.protobuf.ByteString.EMPTY; /** - * optional bytes vertexBinary = 2; - * *
        * SignableVertexSpec
        * 
+ * + * optional bytes vertexBinary = 2; + * @return Whether the vertexBinary field is set. */ + @java.lang.Override public boolean hasVertexBinary() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** - * optional bytes vertexBinary = 2; - * *
        * SignableVertexSpec
        * 
+ * + * optional bytes vertexBinary = 2; + * @return The vertexBinary. */ + @java.lang.Override public com.google.protobuf.ByteString getVertexBinary() { return vertexBinary_; } /** - * optional bytes vertexBinary = 2; - * *
        * SignableVertexSpec
        * 
+ * + * optional bytes vertexBinary = 2; + * @param value The vertexBinary to set. + * @return This builder for chaining. */ public Builder setVertexBinary(com.google.protobuf.ByteString value) { if (value == null) { @@ -7037,11 +7810,12 @@ public Builder setVertexBinary(com.google.protobuf.ByteString value) { return this; } /** - * optional bytes vertexBinary = 2; - * *
        * SignableVertexSpec
        * 
+ * + * optional bytes vertexBinary = 2; + * @return This builder for chaining. */ public Builder clearVertexBinary() { bitField0_ = (bitField0_ & ~0x00000002); @@ -7049,78 +7823,126 @@ public Builder clearVertexBinary() { onChanged(); return this; } + @java.lang.Override + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:VertexOrBinary) } + // @@protoc_insertion_point(class_scope:VertexOrBinary) + private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary DEFAULT_INSTANCE; static { - defaultInstance = new VertexOrBinary(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary(); + } + + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser + PARSER = new com.google.protobuf.AbstractParser() { + @java.lang.Override + public VertexOrBinary parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new VertexOrBinary(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary getDefaultInstanceForType() { + return DEFAULT_INSTANCE; } - // @@protoc_insertion_point(class_scope:VertexOrBinary) } - public interface FragmentRuntimeInfoOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface FragmentRuntimeInfoOrBuilder extends + // @@protoc_insertion_point(interface_extends:FragmentRuntimeInfo) + com.google.protobuf.MessageOrBuilder { - // optional int32 num_self_and_upstream_tasks = 1; /** * optional int32 num_self_and_upstream_tasks = 1; + * @return Whether the numSelfAndUpstreamTasks field is set. */ boolean hasNumSelfAndUpstreamTasks(); /** * optional int32 num_self_and_upstream_tasks = 1; + * @return The numSelfAndUpstreamTasks. */ int getNumSelfAndUpstreamTasks(); - // optional int32 num_self_and_upstream_completed_tasks = 2; /** * optional int32 num_self_and_upstream_completed_tasks = 2; + * @return Whether the numSelfAndUpstreamCompletedTasks field is set. */ boolean hasNumSelfAndUpstreamCompletedTasks(); /** * optional int32 num_self_and_upstream_completed_tasks = 2; + * @return The numSelfAndUpstreamCompletedTasks. */ int getNumSelfAndUpstreamCompletedTasks(); - // optional int32 within_dag_priority = 3; /** * optional int32 within_dag_priority = 3; + * @return Whether the withinDagPriority field is set. */ boolean hasWithinDagPriority(); /** * optional int32 within_dag_priority = 3; + * @return The withinDagPriority. */ int getWithinDagPriority(); - // optional int64 dag_start_time = 4; /** * optional int64 dag_start_time = 4; + * @return Whether the dagStartTime field is set. */ boolean hasDagStartTime(); /** * optional int64 dag_start_time = 4; + * @return The dagStartTime. */ long getDagStartTime(); - // optional int64 first_attempt_start_time = 5; /** * optional int64 first_attempt_start_time = 5; + * @return Whether the firstAttemptStartTime field is set. */ boolean hasFirstAttemptStartTime(); /** * optional int64 first_attempt_start_time = 5; + * @return The firstAttemptStartTime. */ long getFirstAttemptStartTime(); - // optional int64 current_attempt_start_time = 6; /** * optional int64 current_attempt_start_time = 6; + * @return Whether the currentAttemptStartTime field is set. */ boolean hasCurrentAttemptStartTime(); /** * optional int64 current_attempt_start_time = 6; + * @return The currentAttemptStartTime. */ long getCurrentAttemptStartTime(); } @@ -7128,35 +7950,37 @@ public interface FragmentRuntimeInfoOrBuilder * Protobuf type {@code FragmentRuntimeInfo} */ public static final class FragmentRuntimeInfo extends - com.google.protobuf.GeneratedMessage - implements FragmentRuntimeInfoOrBuilder { + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:FragmentRuntimeInfo) + FragmentRuntimeInfoOrBuilder { + private static final long serialVersionUID = 0L; // Use FragmentRuntimeInfo.newBuilder() to construct. - private FragmentRuntimeInfo(com.google.protobuf.GeneratedMessage.Builder builder) { + private FragmentRuntimeInfo(com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private FragmentRuntimeInfo(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final FragmentRuntimeInfo defaultInstance; - public static FragmentRuntimeInfo getDefaultInstance() { - return defaultInstance; + private FragmentRuntimeInfo() { } - public FragmentRuntimeInfo getDefaultInstanceForType() { - return defaultInstance; + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance( + UnusedPrivateParameter unused) { + return new FragmentRuntimeInfo(); } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private FragmentRuntimeInfo( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); @@ -7168,13 +7992,6 @@ private FragmentRuntimeInfo( case 0: done = true; break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } case 8: { bitField0_ |= 0x00000001; numSelfAndUpstreamTasks_ = input.readInt32(); @@ -7205,13 +8022,20 @@ private FragmentRuntimeInfo( currentAttemptStartTime_ = input.readInt64(); break; } + default: { + if (!parseUnknownField( + input, unknownFields, extensionRegistry, tag)) { + done = true; + } + break; + } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); @@ -7222,208 +8046,199 @@ private FragmentRuntimeInfo( return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_FragmentRuntimeInfo_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_FragmentRuntimeInfo_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public FragmentRuntimeInfo parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new FragmentRuntimeInfo(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - private int bitField0_; - // optional int32 num_self_and_upstream_tasks = 1; public static final int NUM_SELF_AND_UPSTREAM_TASKS_FIELD_NUMBER = 1; private int numSelfAndUpstreamTasks_; /** * optional int32 num_self_and_upstream_tasks = 1; + * @return Whether the numSelfAndUpstreamTasks field is set. */ + @java.lang.Override public boolean hasNumSelfAndUpstreamTasks() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional int32 num_self_and_upstream_tasks = 1; + * @return The numSelfAndUpstreamTasks. */ + @java.lang.Override public int getNumSelfAndUpstreamTasks() { return numSelfAndUpstreamTasks_; } - // optional int32 num_self_and_upstream_completed_tasks = 2; public static final int NUM_SELF_AND_UPSTREAM_COMPLETED_TASKS_FIELD_NUMBER = 2; private int numSelfAndUpstreamCompletedTasks_; /** * optional int32 num_self_and_upstream_completed_tasks = 2; + * @return Whether the numSelfAndUpstreamCompletedTasks field is set. */ + @java.lang.Override public boolean hasNumSelfAndUpstreamCompletedTasks() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * optional int32 num_self_and_upstream_completed_tasks = 2; + * @return The numSelfAndUpstreamCompletedTasks. */ + @java.lang.Override public int getNumSelfAndUpstreamCompletedTasks() { return numSelfAndUpstreamCompletedTasks_; } - // optional int32 within_dag_priority = 3; public static final int WITHIN_DAG_PRIORITY_FIELD_NUMBER = 3; private int withinDagPriority_; /** * optional int32 within_dag_priority = 3; + * @return Whether the withinDagPriority field is set. */ + @java.lang.Override public boolean hasWithinDagPriority() { - return ((bitField0_ & 0x00000004) == 0x00000004); + return ((bitField0_ & 0x00000004) != 0); } /** * optional int32 within_dag_priority = 3; + * @return The withinDagPriority. */ + @java.lang.Override public int getWithinDagPriority() { return withinDagPriority_; } - // optional int64 dag_start_time = 4; public static final int DAG_START_TIME_FIELD_NUMBER = 4; private long dagStartTime_; /** * optional int64 dag_start_time = 4; + * @return Whether the dagStartTime field is set. */ + @java.lang.Override public boolean hasDagStartTime() { - return ((bitField0_ & 0x00000008) == 0x00000008); + return ((bitField0_ & 0x00000008) != 0); } /** * optional int64 dag_start_time = 4; + * @return The dagStartTime. */ + @java.lang.Override public long getDagStartTime() { return dagStartTime_; } - // optional int64 first_attempt_start_time = 5; public static final int FIRST_ATTEMPT_START_TIME_FIELD_NUMBER = 5; private long firstAttemptStartTime_; /** * optional int64 first_attempt_start_time = 5; + * @return Whether the firstAttemptStartTime field is set. */ + @java.lang.Override public boolean hasFirstAttemptStartTime() { - return ((bitField0_ & 0x00000010) == 0x00000010); + return ((bitField0_ & 0x00000010) != 0); } /** * optional int64 first_attempt_start_time = 5; + * @return The firstAttemptStartTime. */ + @java.lang.Override public long getFirstAttemptStartTime() { return firstAttemptStartTime_; } - // optional int64 current_attempt_start_time = 6; public static final int CURRENT_ATTEMPT_START_TIME_FIELD_NUMBER = 6; private long currentAttemptStartTime_; /** * optional int64 current_attempt_start_time = 6; + * @return Whether the currentAttemptStartTime field is set. */ + @java.lang.Override public boolean hasCurrentAttemptStartTime() { - return ((bitField0_ & 0x00000020) == 0x00000020); + return ((bitField0_ & 0x00000020) != 0); } /** * optional int64 current_attempt_start_time = 6; + * @return The currentAttemptStartTime. */ + @java.lang.Override public long getCurrentAttemptStartTime() { return currentAttemptStartTime_; } - private void initFields() { - numSelfAndUpstreamTasks_ = 0; - numSelfAndUpstreamCompletedTasks_ = 0; - withinDagPriority_ = 0; - dagStartTime_ = 0L; - firstAttemptStartTime_ = 0L; - currentAttemptStartTime_ = 0L; - } private byte memoizedIsInitialized = -1; + @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } + @java.lang.Override public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); - if (((bitField0_ & 0x00000001) == 0x00000001)) { + if (((bitField0_ & 0x00000001) != 0)) { output.writeInt32(1, numSelfAndUpstreamTasks_); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { + if (((bitField0_ & 0x00000002) != 0)) { output.writeInt32(2, numSelfAndUpstreamCompletedTasks_); } - if (((bitField0_ & 0x00000004) == 0x00000004)) { + if (((bitField0_ & 0x00000004) != 0)) { output.writeInt32(3, withinDagPriority_); } - if (((bitField0_ & 0x00000008) == 0x00000008)) { + if (((bitField0_ & 0x00000008) != 0)) { output.writeInt64(4, dagStartTime_); } - if (((bitField0_ & 0x00000010) == 0x00000010)) { + if (((bitField0_ & 0x00000010) != 0)) { output.writeInt64(5, firstAttemptStartTime_); } - if (((bitField0_ & 0x00000020) == 0x00000020)) { + if (((bitField0_ & 0x00000020) != 0)) { output.writeInt64(6, currentAttemptStartTime_); } - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; + @java.lang.Override public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { + if (((bitField0_ & 0x00000001) != 0)) { size += com.google.protobuf.CodedOutputStream .computeInt32Size(1, numSelfAndUpstreamTasks_); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { + if (((bitField0_ & 0x00000002) != 0)) { size += com.google.protobuf.CodedOutputStream .computeInt32Size(2, numSelfAndUpstreamCompletedTasks_); } - if (((bitField0_ & 0x00000004) == 0x00000004)) { + if (((bitField0_ & 0x00000004) != 0)) { size += com.google.protobuf.CodedOutputStream .computeInt32Size(3, withinDagPriority_); } - if (((bitField0_ & 0x00000008) == 0x00000008)) { + if (((bitField0_ & 0x00000008) != 0)) { size += com.google.protobuf.CodedOutputStream .computeInt64Size(4, dagStartTime_); } - if (((bitField0_ & 0x00000010) == 0x00000010)) { + if (((bitField0_ & 0x00000010) != 0)) { size += com.google.protobuf.CodedOutputStream .computeInt64Size(5, firstAttemptStartTime_); } - if (((bitField0_ & 0x00000020) == 0x00000020)) { + if (((bitField0_ & 0x00000020) != 0)) { size += com.google.protobuf.CodedOutputStream .computeInt64Size(6, currentAttemptStartTime_); } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { @@ -7434,50 +8249,47 @@ public boolean equals(final java.lang.Object obj) { } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo) obj; - boolean result = true; - result = result && (hasNumSelfAndUpstreamTasks() == other.hasNumSelfAndUpstreamTasks()); + if (hasNumSelfAndUpstreamTasks() != other.hasNumSelfAndUpstreamTasks()) return false; if (hasNumSelfAndUpstreamTasks()) { - result = result && (getNumSelfAndUpstreamTasks() - == other.getNumSelfAndUpstreamTasks()); + if (getNumSelfAndUpstreamTasks() + != other.getNumSelfAndUpstreamTasks()) return false; } - result = result && (hasNumSelfAndUpstreamCompletedTasks() == other.hasNumSelfAndUpstreamCompletedTasks()); + if (hasNumSelfAndUpstreamCompletedTasks() != other.hasNumSelfAndUpstreamCompletedTasks()) return false; if (hasNumSelfAndUpstreamCompletedTasks()) { - result = result && (getNumSelfAndUpstreamCompletedTasks() - == other.getNumSelfAndUpstreamCompletedTasks()); + if (getNumSelfAndUpstreamCompletedTasks() + != other.getNumSelfAndUpstreamCompletedTasks()) return false; } - result = result && (hasWithinDagPriority() == other.hasWithinDagPriority()); + if (hasWithinDagPriority() != other.hasWithinDagPriority()) return false; if (hasWithinDagPriority()) { - result = result && (getWithinDagPriority() - == other.getWithinDagPriority()); + if (getWithinDagPriority() + != other.getWithinDagPriority()) return false; } - result = result && (hasDagStartTime() == other.hasDagStartTime()); + if (hasDagStartTime() != other.hasDagStartTime()) return false; if (hasDagStartTime()) { - result = result && (getDagStartTime() - == other.getDagStartTime()); + if (getDagStartTime() + != other.getDagStartTime()) return false; } - result = result && (hasFirstAttemptStartTime() == other.hasFirstAttemptStartTime()); + if (hasFirstAttemptStartTime() != other.hasFirstAttemptStartTime()) return false; if (hasFirstAttemptStartTime()) { - result = result && (getFirstAttemptStartTime() - == other.getFirstAttemptStartTime()); + if (getFirstAttemptStartTime() + != other.getFirstAttemptStartTime()) return false; } - result = result && (hasCurrentAttemptStartTime() == other.hasCurrentAttemptStartTime()); + if (hasCurrentAttemptStartTime() != other.hasCurrentAttemptStartTime()) return false; if (hasCurrentAttemptStartTime()) { - result = result && (getCurrentAttemptStartTime() - == other.getCurrentAttemptStartTime()); + if (getCurrentAttemptStartTime() + != other.getCurrentAttemptStartTime()) return false; } - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; + if (!unknownFields.equals(other.unknownFields)) return false; + return true; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); + hash = (19 * hash) + getDescriptor().hashCode(); if (hasNumSelfAndUpstreamTasks()) { hash = (37 * hash) + NUM_SELF_AND_UPSTREAM_TASKS_FIELD_NUMBER; hash = (53 * hash) + getNumSelfAndUpstreamTasks(); @@ -7492,21 +8304,35 @@ public int hashCode() { } if (hasDagStartTime()) { hash = (37 * hash) + DAG_START_TIME_FIELD_NUMBER; - hash = (53 * hash) + hashLong(getDagStartTime()); + hash = (53 * hash) + com.google.protobuf.Internal.hashLong( + getDagStartTime()); } if (hasFirstAttemptStartTime()) { hash = (37 * hash) + FIRST_ATTEMPT_START_TIME_FIELD_NUMBER; - hash = (53 * hash) + hashLong(getFirstAttemptStartTime()); + hash = (53 * hash) + com.google.protobuf.Internal.hashLong( + getFirstAttemptStartTime()); } if (hasCurrentAttemptStartTime()) { hash = (37 * hash) + CURRENT_ATTEMPT_START_TIME_FIELD_NUMBER; - hash = (53 * hash) + hashLong(getCurrentAttemptStartTime()); + hash = (53 * hash) + com.google.protobuf.Internal.hashLong( + getCurrentAttemptStartTime()); } - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo parseFrom( + java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo parseFrom( + java.nio.ByteBuffer data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { @@ -7530,46 +8356,59 @@ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.Fr } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } + @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -7577,14 +8416,16 @@ protected Builder newBuilderForType( * Protobuf type {@code FragmentRuntimeInfo} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfoOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder implements + // @@protoc_insertion_point(builder_implements:FragmentRuntimeInfo) + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_FragmentRuntimeInfo_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_FragmentRuntimeInfo_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -7597,18 +8438,16 @@ private Builder() { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { } } - private static Builder create() { - return new Builder(); - } - + @java.lang.Override public Builder clear() { super.clear(); numSelfAndUpstreamTasks_ = 0; @@ -7626,19 +8465,18 @@ public Builder clear() { return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - + @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_FragmentRuntimeInfo_descriptor; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo.getDefaultInstance(); } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo result = buildPartial(); if (!result.isInitialized()) { @@ -7647,39 +8485,73 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentR return result; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { + if (((from_bitField0_ & 0x00000001) != 0)) { + result.numSelfAndUpstreamTasks_ = numSelfAndUpstreamTasks_; to_bitField0_ |= 0x00000001; } - result.numSelfAndUpstreamTasks_ = numSelfAndUpstreamTasks_; - if (((from_bitField0_ & 0x00000002) == 0x00000002)) { + if (((from_bitField0_ & 0x00000002) != 0)) { + result.numSelfAndUpstreamCompletedTasks_ = numSelfAndUpstreamCompletedTasks_; to_bitField0_ |= 0x00000002; } - result.numSelfAndUpstreamCompletedTasks_ = numSelfAndUpstreamCompletedTasks_; - if (((from_bitField0_ & 0x00000004) == 0x00000004)) { + if (((from_bitField0_ & 0x00000004) != 0)) { + result.withinDagPriority_ = withinDagPriority_; to_bitField0_ |= 0x00000004; } - result.withinDagPriority_ = withinDagPriority_; - if (((from_bitField0_ & 0x00000008) == 0x00000008)) { + if (((from_bitField0_ & 0x00000008) != 0)) { + result.dagStartTime_ = dagStartTime_; to_bitField0_ |= 0x00000008; } - result.dagStartTime_ = dagStartTime_; - if (((from_bitField0_ & 0x00000010) == 0x00000010)) { + if (((from_bitField0_ & 0x00000010) != 0)) { + result.firstAttemptStartTime_ = firstAttemptStartTime_; to_bitField0_ |= 0x00000010; } - result.firstAttemptStartTime_ = firstAttemptStartTime_; - if (((from_bitField0_ & 0x00000020) == 0x00000020)) { + if (((from_bitField0_ & 0x00000020) != 0)) { + result.currentAttemptStartTime_ = currentAttemptStartTime_; to_bitField0_ |= 0x00000020; } - result.currentAttemptStartTime_ = currentAttemptStartTime_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } + @java.lang.Override + public Builder clone() { + return super.clone(); + } + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.setField(field, value); + } + @java.lang.Override + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + @java.lang.Override + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.addRepeatedField(field, value); + } + @java.lang.Override public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo)other); @@ -7709,14 +8581,17 @@ public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtoc if (other.hasCurrentAttemptStartTime()) { setCurrentAttemptStartTime(other.getCurrentAttemptStartTime()); } - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } + @java.lang.Override public final boolean isInitialized() { return true; } + @java.lang.Override public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) @@ -7726,7 +8601,7 @@ public Builder mergeFrom( parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -7736,22 +8611,27 @@ public Builder mergeFrom( } private int bitField0_; - // optional int32 num_self_and_upstream_tasks = 1; private int numSelfAndUpstreamTasks_ ; /** * optional int32 num_self_and_upstream_tasks = 1; + * @return Whether the numSelfAndUpstreamTasks field is set. */ + @java.lang.Override public boolean hasNumSelfAndUpstreamTasks() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional int32 num_self_and_upstream_tasks = 1; + * @return The numSelfAndUpstreamTasks. */ + @java.lang.Override public int getNumSelfAndUpstreamTasks() { return numSelfAndUpstreamTasks_; } /** * optional int32 num_self_and_upstream_tasks = 1; + * @param value The numSelfAndUpstreamTasks to set. + * @return This builder for chaining. */ public Builder setNumSelfAndUpstreamTasks(int value) { bitField0_ |= 0x00000001; @@ -7761,6 +8641,7 @@ public Builder setNumSelfAndUpstreamTasks(int value) { } /** * optional int32 num_self_and_upstream_tasks = 1; + * @return This builder for chaining. */ public Builder clearNumSelfAndUpstreamTasks() { bitField0_ = (bitField0_ & ~0x00000001); @@ -7769,22 +8650,27 @@ public Builder clearNumSelfAndUpstreamTasks() { return this; } - // optional int32 num_self_and_upstream_completed_tasks = 2; private int numSelfAndUpstreamCompletedTasks_ ; /** * optional int32 num_self_and_upstream_completed_tasks = 2; + * @return Whether the numSelfAndUpstreamCompletedTasks field is set. */ + @java.lang.Override public boolean hasNumSelfAndUpstreamCompletedTasks() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * optional int32 num_self_and_upstream_completed_tasks = 2; + * @return The numSelfAndUpstreamCompletedTasks. */ + @java.lang.Override public int getNumSelfAndUpstreamCompletedTasks() { return numSelfAndUpstreamCompletedTasks_; } /** * optional int32 num_self_and_upstream_completed_tasks = 2; + * @param value The numSelfAndUpstreamCompletedTasks to set. + * @return This builder for chaining. */ public Builder setNumSelfAndUpstreamCompletedTasks(int value) { bitField0_ |= 0x00000002; @@ -7794,6 +8680,7 @@ public Builder setNumSelfAndUpstreamCompletedTasks(int value) { } /** * optional int32 num_self_and_upstream_completed_tasks = 2; + * @return This builder for chaining. */ public Builder clearNumSelfAndUpstreamCompletedTasks() { bitField0_ = (bitField0_ & ~0x00000002); @@ -7802,22 +8689,27 @@ public Builder clearNumSelfAndUpstreamCompletedTasks() { return this; } - // optional int32 within_dag_priority = 3; private int withinDagPriority_ ; /** * optional int32 within_dag_priority = 3; + * @return Whether the withinDagPriority field is set. */ + @java.lang.Override public boolean hasWithinDagPriority() { - return ((bitField0_ & 0x00000004) == 0x00000004); + return ((bitField0_ & 0x00000004) != 0); } /** * optional int32 within_dag_priority = 3; + * @return The withinDagPriority. */ + @java.lang.Override public int getWithinDagPriority() { return withinDagPriority_; } /** * optional int32 within_dag_priority = 3; + * @param value The withinDagPriority to set. + * @return This builder for chaining. */ public Builder setWithinDagPriority(int value) { bitField0_ |= 0x00000004; @@ -7827,6 +8719,7 @@ public Builder setWithinDagPriority(int value) { } /** * optional int32 within_dag_priority = 3; + * @return This builder for chaining. */ public Builder clearWithinDagPriority() { bitField0_ = (bitField0_ & ~0x00000004); @@ -7835,22 +8728,27 @@ public Builder clearWithinDagPriority() { return this; } - // optional int64 dag_start_time = 4; private long dagStartTime_ ; /** * optional int64 dag_start_time = 4; + * @return Whether the dagStartTime field is set. */ + @java.lang.Override public boolean hasDagStartTime() { - return ((bitField0_ & 0x00000008) == 0x00000008); + return ((bitField0_ & 0x00000008) != 0); } /** * optional int64 dag_start_time = 4; + * @return The dagStartTime. */ + @java.lang.Override public long getDagStartTime() { return dagStartTime_; } /** * optional int64 dag_start_time = 4; + * @param value The dagStartTime to set. + * @return This builder for chaining. */ public Builder setDagStartTime(long value) { bitField0_ |= 0x00000008; @@ -7860,6 +8758,7 @@ public Builder setDagStartTime(long value) { } /** * optional int64 dag_start_time = 4; + * @return This builder for chaining. */ public Builder clearDagStartTime() { bitField0_ = (bitField0_ & ~0x00000008); @@ -7868,22 +8767,27 @@ public Builder clearDagStartTime() { return this; } - // optional int64 first_attempt_start_time = 5; private long firstAttemptStartTime_ ; /** * optional int64 first_attempt_start_time = 5; + * @return Whether the firstAttemptStartTime field is set. */ + @java.lang.Override public boolean hasFirstAttemptStartTime() { - return ((bitField0_ & 0x00000010) == 0x00000010); + return ((bitField0_ & 0x00000010) != 0); } /** * optional int64 first_attempt_start_time = 5; + * @return The firstAttemptStartTime. */ + @java.lang.Override public long getFirstAttemptStartTime() { return firstAttemptStartTime_; } /** * optional int64 first_attempt_start_time = 5; + * @param value The firstAttemptStartTime to set. + * @return This builder for chaining. */ public Builder setFirstAttemptStartTime(long value) { bitField0_ |= 0x00000010; @@ -7893,6 +8797,7 @@ public Builder setFirstAttemptStartTime(long value) { } /** * optional int64 first_attempt_start_time = 5; + * @return This builder for chaining. */ public Builder clearFirstAttemptStartTime() { bitField0_ = (bitField0_ & ~0x00000010); @@ -7901,22 +8806,27 @@ public Builder clearFirstAttemptStartTime() { return this; } - // optional int64 current_attempt_start_time = 6; private long currentAttemptStartTime_ ; /** * optional int64 current_attempt_start_time = 6; + * @return Whether the currentAttemptStartTime field is set. */ + @java.lang.Override public boolean hasCurrentAttemptStartTime() { - return ((bitField0_ & 0x00000020) == 0x00000020); + return ((bitField0_ & 0x00000020) != 0); } /** * optional int64 current_attempt_start_time = 6; + * @return The currentAttemptStartTime. */ + @java.lang.Override public long getCurrentAttemptStartTime() { return currentAttemptStartTime_; } /** * optional int64 current_attempt_start_time = 6; + * @param value The currentAttemptStartTime to set. + * @return This builder for chaining. */ public Builder setCurrentAttemptStartTime(long value) { bitField0_ |= 0x00000020; @@ -7926,6 +8836,7 @@ public Builder setCurrentAttemptStartTime(long value) { } /** * optional int64 current_attempt_start_time = 6; + * @return This builder for chaining. */ public Builder clearCurrentAttemptStartTime() { bitField0_ = (bitField0_ & ~0x00000020); @@ -7933,53 +8844,99 @@ public Builder clearCurrentAttemptStartTime() { onChanged(); return this; } + @java.lang.Override + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:FragmentRuntimeInfo) } + // @@protoc_insertion_point(class_scope:FragmentRuntimeInfo) + private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo DEFAULT_INSTANCE; static { - defaultInstance = new FragmentRuntimeInfo(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo(); + } + + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser + PARSER = new com.google.protobuf.AbstractParser() { + @java.lang.Override + public FragmentRuntimeInfo parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new FragmentRuntimeInfo(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo getDefaultInstanceForType() { + return DEFAULT_INSTANCE; } - // @@protoc_insertion_point(class_scope:FragmentRuntimeInfo) } - public interface QueryIdentifierProtoOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface QueryIdentifierProtoOrBuilder extends + // @@protoc_insertion_point(interface_extends:QueryIdentifierProto) + com.google.protobuf.MessageOrBuilder { - // optional string application_id_string = 1; /** * optional string application_id_string = 1; + * @return Whether the applicationIdString field is set. */ boolean hasApplicationIdString(); /** * optional string application_id_string = 1; + * @return The applicationIdString. */ java.lang.String getApplicationIdString(); /** * optional string application_id_string = 1; + * @return The bytes for applicationIdString. */ com.google.protobuf.ByteString getApplicationIdStringBytes(); - // optional int32 dag_index = 2; /** * optional int32 dag_index = 2; + * @return Whether the dagIndex field is set. */ boolean hasDagIndex(); /** * optional int32 dag_index = 2; + * @return The dagIndex. */ int getDagIndex(); - // optional int32 app_attempt_number = 3; /** * optional int32 app_attempt_number = 3; + * @return Whether the appAttemptNumber field is set. */ boolean hasAppAttemptNumber(); /** * optional int32 app_attempt_number = 3; + * @return The appAttemptNumber. */ int getAppAttemptNumber(); } @@ -7987,35 +8944,38 @@ public interface QueryIdentifierProtoOrBuilder * Protobuf type {@code QueryIdentifierProto} */ public static final class QueryIdentifierProto extends - com.google.protobuf.GeneratedMessage - implements QueryIdentifierProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:QueryIdentifierProto) + QueryIdentifierProtoOrBuilder { + private static final long serialVersionUID = 0L; // Use QueryIdentifierProto.newBuilder() to construct. - private QueryIdentifierProto(com.google.protobuf.GeneratedMessage.Builder builder) { + private QueryIdentifierProto(com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private QueryIdentifierProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final QueryIdentifierProto defaultInstance; - public static QueryIdentifierProto getDefaultInstance() { - return defaultInstance; + private QueryIdentifierProto() { + applicationIdString_ = ""; } - public QueryIdentifierProto getDefaultInstanceForType() { - return defaultInstance; + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance( + UnusedPrivateParameter unused) { + return new QueryIdentifierProto(); } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private QueryIdentifierProto( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); @@ -8027,16 +8987,10 @@ private QueryIdentifierProto( case 0: done = true; break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } case 10: { + com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000001; - applicationIdString_ = input.readBytes(); + applicationIdString_ = bs; break; } case 16: { @@ -8049,13 +9003,20 @@ private QueryIdentifierProto( appAttemptNumber_ = input.readInt32(); break; } + default: { + if (!parseUnknownField( + input, unknownFields, extensionRegistry, tag)) { + done = true; + } + break; + } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); @@ -8066,41 +9027,30 @@ private QueryIdentifierProto( return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryIdentifierProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryIdentifierProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public QueryIdentifierProto parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new QueryIdentifierProto(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - private int bitField0_; - // optional string application_id_string = 1; public static final int APPLICATION_ID_STRING_FIELD_NUMBER = 1; - private java.lang.Object applicationIdString_; + private volatile java.lang.Object applicationIdString_; /** * optional string application_id_string = 1; + * @return Whether the applicationIdString field is set. */ + @java.lang.Override public boolean hasApplicationIdString() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional string application_id_string = 1; + * @return The applicationIdString. */ + @java.lang.Override public java.lang.String getApplicationIdString() { java.lang.Object ref = applicationIdString_; if (ref instanceof java.lang.String) { @@ -8117,7 +9067,9 @@ public java.lang.String getApplicationIdString() { } /** * optional string application_id_string = 1; + * @return The bytes for applicationIdString. */ + @java.lang.Override public com.google.protobuf.ByteString getApplicationIdStringBytes() { java.lang.Object ref = applicationIdString_; @@ -8132,97 +9084,92 @@ public java.lang.String getApplicationIdString() { } } - // optional int32 dag_index = 2; public static final int DAG_INDEX_FIELD_NUMBER = 2; private int dagIndex_; /** * optional int32 dag_index = 2; + * @return Whether the dagIndex field is set. */ + @java.lang.Override public boolean hasDagIndex() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * optional int32 dag_index = 2; + * @return The dagIndex. */ + @java.lang.Override public int getDagIndex() { return dagIndex_; } - // optional int32 app_attempt_number = 3; public static final int APP_ATTEMPT_NUMBER_FIELD_NUMBER = 3; private int appAttemptNumber_; /** * optional int32 app_attempt_number = 3; + * @return Whether the appAttemptNumber field is set. */ + @java.lang.Override public boolean hasAppAttemptNumber() { - return ((bitField0_ & 0x00000004) == 0x00000004); + return ((bitField0_ & 0x00000004) != 0); } /** * optional int32 app_attempt_number = 3; + * @return The appAttemptNumber. */ + @java.lang.Override public int getAppAttemptNumber() { return appAttemptNumber_; } - private void initFields() { - applicationIdString_ = ""; - dagIndex_ = 0; - appAttemptNumber_ = 0; - } private byte memoizedIsInitialized = -1; + @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } + @java.lang.Override public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); - if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeBytes(1, getApplicationIdStringBytes()); + if (((bitField0_ & 0x00000001) != 0)) { + com.google.protobuf.GeneratedMessageV3.writeString(output, 1, applicationIdString_); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { + if (((bitField0_ & 0x00000002) != 0)) { output.writeInt32(2, dagIndex_); } - if (((bitField0_ & 0x00000004) == 0x00000004)) { + if (((bitField0_ & 0x00000004) != 0)) { output.writeInt32(3, appAttemptNumber_); } - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; + @java.lang.Override public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(1, getApplicationIdStringBytes()); + if (((bitField0_ & 0x00000001) != 0)) { + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, applicationIdString_); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { + if (((bitField0_ & 0x00000002) != 0)) { size += com.google.protobuf.CodedOutputStream .computeInt32Size(2, dagIndex_); } - if (((bitField0_ & 0x00000004) == 0x00000004)) { + if (((bitField0_ & 0x00000004) != 0)) { size += com.google.protobuf.CodedOutputStream .computeInt32Size(3, appAttemptNumber_); } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { @@ -8233,35 +9180,32 @@ public boolean equals(final java.lang.Object obj) { } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto) obj; - boolean result = true; - result = result && (hasApplicationIdString() == other.hasApplicationIdString()); + if (hasApplicationIdString() != other.hasApplicationIdString()) return false; if (hasApplicationIdString()) { - result = result && getApplicationIdString() - .equals(other.getApplicationIdString()); + if (!getApplicationIdString() + .equals(other.getApplicationIdString())) return false; } - result = result && (hasDagIndex() == other.hasDagIndex()); + if (hasDagIndex() != other.hasDagIndex()) return false; if (hasDagIndex()) { - result = result && (getDagIndex() - == other.getDagIndex()); + if (getDagIndex() + != other.getDagIndex()) return false; } - result = result && (hasAppAttemptNumber() == other.hasAppAttemptNumber()); + if (hasAppAttemptNumber() != other.hasAppAttemptNumber()) return false; if (hasAppAttemptNumber()) { - result = result && (getAppAttemptNumber() - == other.getAppAttemptNumber()); + if (getAppAttemptNumber() + != other.getAppAttemptNumber()) return false; } - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; + if (!unknownFields.equals(other.unknownFields)) return false; + return true; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); + hash = (19 * hash) + getDescriptor().hashCode(); if (hasApplicationIdString()) { hash = (37 * hash) + APPLICATION_ID_STRING_FIELD_NUMBER; hash = (53 * hash) + getApplicationIdString().hashCode(); @@ -8274,11 +9218,22 @@ public int hashCode() { hash = (37 * hash) + APP_ATTEMPT_NUMBER_FIELD_NUMBER; hash = (53 * hash) + getAppAttemptNumber(); } - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parseFrom( + java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parseFrom( + java.nio.ByteBuffer data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { @@ -8302,46 +9257,59 @@ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.Qu } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } + @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -8349,14 +9317,16 @@ protected Builder newBuilderForType( * Protobuf type {@code QueryIdentifierProto} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder implements + // @@protoc_insertion_point(builder_implements:QueryIdentifierProto) + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryIdentifierProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryIdentifierProto_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -8369,18 +9339,16 @@ private Builder() { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { } } - private static Builder create() { - return new Builder(); - } - + @java.lang.Override public Builder clear() { super.clear(); applicationIdString_ = ""; @@ -8392,19 +9360,18 @@ public Builder clear() { return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - + @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryIdentifierProto_descriptor; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance(); } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto result = buildPartial(); if (!result.isInitialized()) { @@ -8413,27 +9380,61 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIden return result; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { + if (((from_bitField0_ & 0x00000001) != 0)) { to_bitField0_ |= 0x00000001; } result.applicationIdString_ = applicationIdString_; - if (((from_bitField0_ & 0x00000002) == 0x00000002)) { + if (((from_bitField0_ & 0x00000002) != 0)) { + result.dagIndex_ = dagIndex_; to_bitField0_ |= 0x00000002; } - result.dagIndex_ = dagIndex_; - if (((from_bitField0_ & 0x00000004) == 0x00000004)) { + if (((from_bitField0_ & 0x00000004) != 0)) { + result.appAttemptNumber_ = appAttemptNumber_; to_bitField0_ |= 0x00000004; } - result.appAttemptNumber_ = appAttemptNumber_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } + @java.lang.Override + public Builder clone() { + return super.clone(); + } + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.setField(field, value); + } + @java.lang.Override + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + @java.lang.Override + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.addRepeatedField(field, value); + } + @java.lang.Override public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto)other); @@ -8456,14 +9457,17 @@ public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtoc if (other.hasAppAttemptNumber()) { setAppAttemptNumber(other.getAppAttemptNumber()); } - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } + @java.lang.Override public final boolean isInitialized() { return true; } + @java.lang.Override public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) @@ -8473,7 +9477,7 @@ public Builder mergeFrom( parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -8483,23 +9487,27 @@ public Builder mergeFrom( } private int bitField0_; - // optional string application_id_string = 1; private java.lang.Object applicationIdString_ = ""; /** * optional string application_id_string = 1; + * @return Whether the applicationIdString field is set. */ public boolean hasApplicationIdString() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional string application_id_string = 1; + * @return The applicationIdString. */ public java.lang.String getApplicationIdString() { java.lang.Object ref = applicationIdString_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - applicationIdString_ = s; + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + applicationIdString_ = s; + } return s; } else { return (java.lang.String) ref; @@ -8507,6 +9515,7 @@ public java.lang.String getApplicationIdString() { } /** * optional string application_id_string = 1; + * @return The bytes for applicationIdString. */ public com.google.protobuf.ByteString getApplicationIdStringBytes() { @@ -8523,6 +9532,8 @@ public java.lang.String getApplicationIdString() { } /** * optional string application_id_string = 1; + * @param value The applicationIdString to set. + * @return This builder for chaining. */ public Builder setApplicationIdString( java.lang.String value) { @@ -8536,6 +9547,7 @@ public Builder setApplicationIdString( } /** * optional string application_id_string = 1; + * @return This builder for chaining. */ public Builder clearApplicationIdString() { bitField0_ = (bitField0_ & ~0x00000001); @@ -8545,6 +9557,8 @@ public Builder clearApplicationIdString() { } /** * optional string application_id_string = 1; + * @param value The bytes for applicationIdString to set. + * @return This builder for chaining. */ public Builder setApplicationIdStringBytes( com.google.protobuf.ByteString value) { @@ -8557,22 +9571,27 @@ public Builder setApplicationIdStringBytes( return this; } - // optional int32 dag_index = 2; private int dagIndex_ ; /** * optional int32 dag_index = 2; + * @return Whether the dagIndex field is set. */ + @java.lang.Override public boolean hasDagIndex() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * optional int32 dag_index = 2; + * @return The dagIndex. */ + @java.lang.Override public int getDagIndex() { return dagIndex_; } /** * optional int32 dag_index = 2; + * @param value The dagIndex to set. + * @return This builder for chaining. */ public Builder setDagIndex(int value) { bitField0_ |= 0x00000002; @@ -8582,6 +9601,7 @@ public Builder setDagIndex(int value) { } /** * optional int32 dag_index = 2; + * @return This builder for chaining. */ public Builder clearDagIndex() { bitField0_ = (bitField0_ & ~0x00000002); @@ -8590,22 +9610,27 @@ public Builder clearDagIndex() { return this; } - // optional int32 app_attempt_number = 3; private int appAttemptNumber_ ; /** * optional int32 app_attempt_number = 3; + * @return Whether the appAttemptNumber field is set. */ + @java.lang.Override public boolean hasAppAttemptNumber() { - return ((bitField0_ & 0x00000004) == 0x00000004); + return ((bitField0_ & 0x00000004) != 0); } /** * optional int32 app_attempt_number = 3; + * @return The appAttemptNumber. */ + @java.lang.Override public int getAppAttemptNumber() { return appAttemptNumber_; } /** * optional int32 app_attempt_number = 3; + * @param value The appAttemptNumber to set. + * @return This builder for chaining. */ public Builder setAppAttemptNumber(int value) { bitField0_ |= 0x00000004; @@ -8615,6 +9640,7 @@ public Builder setAppAttemptNumber(int value) { } /** * optional int32 app_attempt_number = 3; + * @return This builder for chaining. */ public Builder clearAppAttemptNumber() { bitField0_ = (bitField0_ & ~0x00000004); @@ -8622,74 +9648,120 @@ public Builder clearAppAttemptNumber() { onChanged(); return this; } + @java.lang.Override + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } - // @@protoc_insertion_point(builder_scope:QueryIdentifierProto) - } + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } - static { - defaultInstance = new QueryIdentifierProto(true); - defaultInstance.initFields(); + + // @@protoc_insertion_point(builder_scope:QueryIdentifierProto) } // @@protoc_insertion_point(class_scope:QueryIdentifierProto) + private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto DEFAULT_INSTANCE; + static { + DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto(); + } + + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser + PARSER = new com.google.protobuf.AbstractParser() { + @java.lang.Override + public QueryIdentifierProto parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new QueryIdentifierProto(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto getDefaultInstanceForType() { + return DEFAULT_INSTANCE; + } + } - public interface NotTezEventOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface NotTezEventOrBuilder extends + // @@protoc_insertion_point(interface_extends:NotTezEvent) + com.google.protobuf.MessageOrBuilder { - // required bytes input_event_proto_bytes = 1; /** * required bytes input_event_proto_bytes = 1; + * @return Whether the inputEventProtoBytes field is set. */ boolean hasInputEventProtoBytes(); /** * required bytes input_event_proto_bytes = 1; + * @return The inputEventProtoBytes. */ com.google.protobuf.ByteString getInputEventProtoBytes(); - // required string vertex_name = 2; /** * required string vertex_name = 2; + * @return Whether the vertexName field is set. */ boolean hasVertexName(); /** * required string vertex_name = 2; + * @return The vertexName. */ java.lang.String getVertexName(); /** * required string vertex_name = 2; + * @return The bytes for vertexName. */ com.google.protobuf.ByteString getVertexNameBytes(); - // required string dest_input_name = 3; /** * required string dest_input_name = 3; + * @return Whether the destInputName field is set. */ boolean hasDestInputName(); /** * required string dest_input_name = 3; + * @return The destInputName. */ java.lang.String getDestInputName(); /** * required string dest_input_name = 3; + * @return The bytes for destInputName. */ com.google.protobuf.ByteString getDestInputNameBytes(); - // optional int32 key_id = 4; /** * optional int32 key_id = 4; + * @return Whether the keyId field is set. */ boolean hasKeyId(); /** * optional int32 key_id = 4; + * @return The keyId. */ int getKeyId(); } /** - * Protobuf type {@code NotTezEvent} - * *
    **
    * Tez API implementation derives an enum value from instanceof on the event, then uses that enum
@@ -8698,37 +9770,44 @@ public interface NotTezEventOrBuilder
    * three times over to add anything there. So, we'd do our own "inspired" serialization.
    * Eventually we'll move away from events for API.
    * 
+ * + * Protobuf type {@code NotTezEvent} */ public static final class NotTezEvent extends - com.google.protobuf.GeneratedMessage - implements NotTezEventOrBuilder { + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:NotTezEvent) + NotTezEventOrBuilder { + private static final long serialVersionUID = 0L; // Use NotTezEvent.newBuilder() to construct. - private NotTezEvent(com.google.protobuf.GeneratedMessage.Builder builder) { + private NotTezEvent(com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private NotTezEvent(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final NotTezEvent defaultInstance; - public static NotTezEvent getDefaultInstance() { - return defaultInstance; + private NotTezEvent() { + inputEventProtoBytes_ = com.google.protobuf.ByteString.EMPTY; + vertexName_ = ""; + destInputName_ = ""; } - public NotTezEvent getDefaultInstanceForType() { - return defaultInstance; + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance( + UnusedPrivateParameter unused) { + return new NotTezEvent(); } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private NotTezEvent( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); @@ -8740,26 +9819,21 @@ private NotTezEvent( case 0: done = true; break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } case 10: { bitField0_ |= 0x00000001; inputEventProtoBytes_ = input.readBytes(); break; } case 18: { + com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000002; - vertexName_ = input.readBytes(); + vertexName_ = bs; break; } case 26: { + com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000004; - destInputName_ = input.readBytes(); + destInputName_ = bs; break; } case 32: { @@ -8767,13 +9841,20 @@ private NotTezEvent( keyId_ = input.readInt32(); break; } + default: { + if (!parseUnknownField( + input, unknownFields, extensionRegistry, tag)) { + done = true; + } + break; + } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); @@ -8784,57 +9865,49 @@ private NotTezEvent( return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_NotTezEvent_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_NotTezEvent_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public NotTezEvent parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new NotTezEvent(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - private int bitField0_; - // required bytes input_event_proto_bytes = 1; public static final int INPUT_EVENT_PROTO_BYTES_FIELD_NUMBER = 1; private com.google.protobuf.ByteString inputEventProtoBytes_; /** * required bytes input_event_proto_bytes = 1; + * @return Whether the inputEventProtoBytes field is set. */ + @java.lang.Override public boolean hasInputEventProtoBytes() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * required bytes input_event_proto_bytes = 1; + * @return The inputEventProtoBytes. */ + @java.lang.Override public com.google.protobuf.ByteString getInputEventProtoBytes() { return inputEventProtoBytes_; } - // required string vertex_name = 2; public static final int VERTEX_NAME_FIELD_NUMBER = 2; - private java.lang.Object vertexName_; + private volatile java.lang.Object vertexName_; /** * required string vertex_name = 2; + * @return Whether the vertexName field is set. */ + @java.lang.Override public boolean hasVertexName() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * required string vertex_name = 2; + * @return The vertexName. */ + @java.lang.Override public java.lang.String getVertexName() { java.lang.Object ref = vertexName_; if (ref instanceof java.lang.String) { @@ -8851,7 +9924,9 @@ public java.lang.String getVertexName() { } /** * required string vertex_name = 2; + * @return The bytes for vertexName. */ + @java.lang.Override public com.google.protobuf.ByteString getVertexNameBytes() { java.lang.Object ref = vertexName_; @@ -8866,18 +9941,21 @@ public java.lang.String getVertexName() { } } - // required string dest_input_name = 3; public static final int DEST_INPUT_NAME_FIELD_NUMBER = 3; - private java.lang.Object destInputName_; + private volatile java.lang.Object destInputName_; /** * required string dest_input_name = 3; + * @return Whether the destInputName field is set. */ + @java.lang.Override public boolean hasDestInputName() { - return ((bitField0_ & 0x00000004) == 0x00000004); + return ((bitField0_ & 0x00000004) != 0); } /** * required string dest_input_name = 3; + * @return The destInputName. */ + @java.lang.Override public java.lang.String getDestInputName() { java.lang.Object ref = destInputName_; if (ref instanceof java.lang.String) { @@ -8894,7 +9972,9 @@ public java.lang.String getDestInputName() { } /** * required string dest_input_name = 3; + * @return The bytes for destInputName. */ + @java.lang.Override public com.google.protobuf.ByteString getDestInputNameBytes() { java.lang.Object ref = destInputName_; @@ -8909,32 +9989,31 @@ public java.lang.String getDestInputName() { } } - // optional int32 key_id = 4; public static final int KEY_ID_FIELD_NUMBER = 4; private int keyId_; /** * optional int32 key_id = 4; + * @return Whether the keyId field is set. */ + @java.lang.Override public boolean hasKeyId() { - return ((bitField0_ & 0x00000008) == 0x00000008); + return ((bitField0_ & 0x00000008) != 0); } /** * optional int32 key_id = 4; + * @return The keyId. */ + @java.lang.Override public int getKeyId() { return keyId_; } - private void initFields() { - inputEventProtoBytes_ = com.google.protobuf.ByteString.EMPTY; - vertexName_ = ""; - destInputName_ = ""; - keyId_ = 0; - } private byte memoizedIsInitialized = -1; + @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; if (!hasInputEventProtoBytes()) { memoizedIsInitialized = 0; @@ -8952,58 +10031,49 @@ public final boolean isInitialized() { return true; } + @java.lang.Override public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); - if (((bitField0_ & 0x00000001) == 0x00000001)) { + if (((bitField0_ & 0x00000001) != 0)) { output.writeBytes(1, inputEventProtoBytes_); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - output.writeBytes(2, getVertexNameBytes()); + if (((bitField0_ & 0x00000002) != 0)) { + com.google.protobuf.GeneratedMessageV3.writeString(output, 2, vertexName_); } - if (((bitField0_ & 0x00000004) == 0x00000004)) { - output.writeBytes(3, getDestInputNameBytes()); + if (((bitField0_ & 0x00000004) != 0)) { + com.google.protobuf.GeneratedMessageV3.writeString(output, 3, destInputName_); } - if (((bitField0_ & 0x00000008) == 0x00000008)) { + if (((bitField0_ & 0x00000008) != 0)) { output.writeInt32(4, keyId_); } - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; + @java.lang.Override public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { + if (((bitField0_ & 0x00000001) != 0)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(1, inputEventProtoBytes_); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(2, getVertexNameBytes()); + if (((bitField0_ & 0x00000002) != 0)) { + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(2, vertexName_); } - if (((bitField0_ & 0x00000004) == 0x00000004)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(3, getDestInputNameBytes()); + if (((bitField0_ & 0x00000004) != 0)) { + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(3, destInputName_); } - if (((bitField0_ & 0x00000008) == 0x00000008)) { + if (((bitField0_ & 0x00000008) != 0)) { size += com.google.protobuf.CodedOutputStream .computeInt32Size(4, keyId_); } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { @@ -9014,40 +10084,37 @@ public boolean equals(final java.lang.Object obj) { } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent) obj; - boolean result = true; - result = result && (hasInputEventProtoBytes() == other.hasInputEventProtoBytes()); + if (hasInputEventProtoBytes() != other.hasInputEventProtoBytes()) return false; if (hasInputEventProtoBytes()) { - result = result && getInputEventProtoBytes() - .equals(other.getInputEventProtoBytes()); + if (!getInputEventProtoBytes() + .equals(other.getInputEventProtoBytes())) return false; } - result = result && (hasVertexName() == other.hasVertexName()); + if (hasVertexName() != other.hasVertexName()) return false; if (hasVertexName()) { - result = result && getVertexName() - .equals(other.getVertexName()); + if (!getVertexName() + .equals(other.getVertexName())) return false; } - result = result && (hasDestInputName() == other.hasDestInputName()); + if (hasDestInputName() != other.hasDestInputName()) return false; if (hasDestInputName()) { - result = result && getDestInputName() - .equals(other.getDestInputName()); + if (!getDestInputName() + .equals(other.getDestInputName())) return false; } - result = result && (hasKeyId() == other.hasKeyId()); + if (hasKeyId() != other.hasKeyId()) return false; if (hasKeyId()) { - result = result && (getKeyId() - == other.getKeyId()); + if (getKeyId() + != other.getKeyId()) return false; } - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; + if (!unknownFields.equals(other.unknownFields)) return false; + return true; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); + hash = (19 * hash) + getDescriptor().hashCode(); if (hasInputEventProtoBytes()) { hash = (37 * hash) + INPUT_EVENT_PROTO_BYTES_FIELD_NUMBER; hash = (53 * hash) + getInputEventProtoBytes().hashCode(); @@ -9064,11 +10131,22 @@ public int hashCode() { hash = (37 * hash) + KEY_ID_FIELD_NUMBER; hash = (53 * hash) + getKeyId(); } - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent parseFrom( + java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent parseFrom( + java.nio.ByteBuffer data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { @@ -9092,52 +10170,63 @@ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.No } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } + @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** - * Protobuf type {@code NotTezEvent} - * *
      **
      * Tez API implementation derives an enum value from instanceof on the event, then uses that enum
@@ -9146,16 +10235,20 @@ protected Builder newBuilderForType(
      * three times over to add anything there. So, we'd do our own "inspired" serialization.
      * Eventually we'll move away from events for API.
      * 
+ * + * Protobuf type {@code NotTezEvent} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEventOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder implements + // @@protoc_insertion_point(builder_implements:NotTezEvent) + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEventOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_NotTezEvent_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_NotTezEvent_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -9168,18 +10261,16 @@ private Builder() { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { } } - private static Builder create() { - return new Builder(); - } - + @java.lang.Override public Builder clear() { super.clear(); inputEventProtoBytes_ = com.google.protobuf.ByteString.EMPTY; @@ -9193,19 +10284,18 @@ public Builder clear() { return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - + @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_NotTezEvent_descriptor; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent.getDefaultInstance(); } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent result = buildPartial(); if (!result.isInitialized()) { @@ -9214,31 +10304,65 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEve return result; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { + if (((from_bitField0_ & 0x00000001) != 0)) { to_bitField0_ |= 0x00000001; } result.inputEventProtoBytes_ = inputEventProtoBytes_; - if (((from_bitField0_ & 0x00000002) == 0x00000002)) { + if (((from_bitField0_ & 0x00000002) != 0)) { to_bitField0_ |= 0x00000002; } result.vertexName_ = vertexName_; - if (((from_bitField0_ & 0x00000004) == 0x00000004)) { + if (((from_bitField0_ & 0x00000004) != 0)) { to_bitField0_ |= 0x00000004; } result.destInputName_ = destInputName_; - if (((from_bitField0_ & 0x00000008) == 0x00000008)) { + if (((from_bitField0_ & 0x00000008) != 0)) { + result.keyId_ = keyId_; to_bitField0_ |= 0x00000008; } - result.keyId_ = keyId_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } + @java.lang.Override + public Builder clone() { + return super.clone(); + } + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.setField(field, value); + } + @java.lang.Override + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + @java.lang.Override + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.addRepeatedField(field, value); + } + @java.lang.Override public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent)other); @@ -9266,26 +10390,26 @@ public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtoc if (other.hasKeyId()) { setKeyId(other.getKeyId()); } - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } + @java.lang.Override public final boolean isInitialized() { if (!hasInputEventProtoBytes()) { - return false; } if (!hasVertexName()) { - return false; } if (!hasDestInputName()) { - return false; } return true; } + @java.lang.Override public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) @@ -9295,7 +10419,7 @@ public Builder mergeFrom( parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -9305,22 +10429,27 @@ public Builder mergeFrom( } private int bitField0_; - // required bytes input_event_proto_bytes = 1; private com.google.protobuf.ByteString inputEventProtoBytes_ = com.google.protobuf.ByteString.EMPTY; /** * required bytes input_event_proto_bytes = 1; + * @return Whether the inputEventProtoBytes field is set. */ + @java.lang.Override public boolean hasInputEventProtoBytes() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * required bytes input_event_proto_bytes = 1; + * @return The inputEventProtoBytes. */ + @java.lang.Override public com.google.protobuf.ByteString getInputEventProtoBytes() { return inputEventProtoBytes_; } /** * required bytes input_event_proto_bytes = 1; + * @param value The inputEventProtoBytes to set. + * @return This builder for chaining. */ public Builder setInputEventProtoBytes(com.google.protobuf.ByteString value) { if (value == null) { @@ -9333,6 +10462,7 @@ public Builder setInputEventProtoBytes(com.google.protobuf.ByteString value) { } /** * required bytes input_event_proto_bytes = 1; + * @return This builder for chaining. */ public Builder clearInputEventProtoBytes() { bitField0_ = (bitField0_ & ~0x00000001); @@ -9341,23 +10471,27 @@ public Builder clearInputEventProtoBytes() { return this; } - // required string vertex_name = 2; private java.lang.Object vertexName_ = ""; /** * required string vertex_name = 2; + * @return Whether the vertexName field is set. */ public boolean hasVertexName() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * required string vertex_name = 2; + * @return The vertexName. */ public java.lang.String getVertexName() { java.lang.Object ref = vertexName_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - vertexName_ = s; + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + vertexName_ = s; + } return s; } else { return (java.lang.String) ref; @@ -9365,6 +10499,7 @@ public java.lang.String getVertexName() { } /** * required string vertex_name = 2; + * @return The bytes for vertexName. */ public com.google.protobuf.ByteString getVertexNameBytes() { @@ -9381,6 +10516,8 @@ public java.lang.String getVertexName() { } /** * required string vertex_name = 2; + * @param value The vertexName to set. + * @return This builder for chaining. */ public Builder setVertexName( java.lang.String value) { @@ -9394,6 +10531,7 @@ public Builder setVertexName( } /** * required string vertex_name = 2; + * @return This builder for chaining. */ public Builder clearVertexName() { bitField0_ = (bitField0_ & ~0x00000002); @@ -9403,6 +10541,8 @@ public Builder clearVertexName() { } /** * required string vertex_name = 2; + * @param value The bytes for vertexName to set. + * @return This builder for chaining. */ public Builder setVertexNameBytes( com.google.protobuf.ByteString value) { @@ -9415,23 +10555,27 @@ public Builder setVertexNameBytes( return this; } - // required string dest_input_name = 3; private java.lang.Object destInputName_ = ""; /** * required string dest_input_name = 3; + * @return Whether the destInputName field is set. */ public boolean hasDestInputName() { - return ((bitField0_ & 0x00000004) == 0x00000004); + return ((bitField0_ & 0x00000004) != 0); } /** * required string dest_input_name = 3; + * @return The destInputName. */ public java.lang.String getDestInputName() { java.lang.Object ref = destInputName_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - destInputName_ = s; + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + destInputName_ = s; + } return s; } else { return (java.lang.String) ref; @@ -9439,6 +10583,7 @@ public java.lang.String getDestInputName() { } /** * required string dest_input_name = 3; + * @return The bytes for destInputName. */ public com.google.protobuf.ByteString getDestInputNameBytes() { @@ -9455,6 +10600,8 @@ public java.lang.String getDestInputName() { } /** * required string dest_input_name = 3; + * @param value The destInputName to set. + * @return This builder for chaining. */ public Builder setDestInputName( java.lang.String value) { @@ -9468,6 +10615,7 @@ public Builder setDestInputName( } /** * required string dest_input_name = 3; + * @return This builder for chaining. */ public Builder clearDestInputName() { bitField0_ = (bitField0_ & ~0x00000004); @@ -9477,6 +10625,8 @@ public Builder clearDestInputName() { } /** * required string dest_input_name = 3; + * @param value The bytes for destInputName to set. + * @return This builder for chaining. */ public Builder setDestInputNameBytes( com.google.protobuf.ByteString value) { @@ -9489,22 +10639,27 @@ public Builder setDestInputNameBytes( return this; } - // optional int32 key_id = 4; private int keyId_ ; /** * optional int32 key_id = 4; + * @return Whether the keyId field is set. */ + @java.lang.Override public boolean hasKeyId() { - return ((bitField0_ & 0x00000008) == 0x00000008); + return ((bitField0_ & 0x00000008) != 0); } /** * optional int32 key_id = 4; + * @return The keyId. */ + @java.lang.Override public int getKeyId() { return keyId_; } /** * optional int32 key_id = 4; + * @param value The keyId to set. + * @return This builder for chaining. */ public Builder setKeyId(int value) { bitField0_ |= 0x00000008; @@ -9514,6 +10669,7 @@ public Builder setKeyId(int value) { } /** * optional int32 key_id = 4; + * @return This builder for chaining. */ public Builder clearKeyId() { bitField0_ = (bitField0_ & ~0x00000008); @@ -9521,28 +10677,71 @@ public Builder clearKeyId() { onChanged(); return this; } + @java.lang.Override + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:NotTezEvent) } + // @@protoc_insertion_point(class_scope:NotTezEvent) + private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent DEFAULT_INSTANCE; static { - defaultInstance = new NotTezEvent(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent(); + } + + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser + PARSER = new com.google.protobuf.AbstractParser() { + @java.lang.Override + public NotTezEvent parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new NotTezEvent(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.NotTezEvent getDefaultInstanceForType() { + return DEFAULT_INSTANCE; } - // @@protoc_insertion_point(class_scope:NotTezEvent) } - public interface SubmitWorkRequestProtoOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface SubmitWorkRequestProtoOrBuilder extends + // @@protoc_insertion_point(interface_extends:SubmitWorkRequestProto) + com.google.protobuf.MessageOrBuilder { - // optional .VertexOrBinary work_spec = 1; /** * optional .VertexOrBinary work_spec = 1; + * @return Whether the workSpec field is set. */ boolean hasWorkSpec(); /** * optional .VertexOrBinary work_spec = 1; + * @return The workSpec. */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary getWorkSpec(); /** @@ -9550,180 +10749,196 @@ public interface SubmitWorkRequestProtoOrBuilder */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinaryOrBuilder getWorkSpecOrBuilder(); - // optional bytes work_spec_signature = 2; /** * optional bytes work_spec_signature = 2; + * @return Whether the workSpecSignature field is set. */ boolean hasWorkSpecSignature(); /** * optional bytes work_spec_signature = 2; + * @return The workSpecSignature. */ com.google.protobuf.ByteString getWorkSpecSignature(); - // optional int32 fragment_number = 3; /** * optional int32 fragment_number = 3; + * @return Whether the fragmentNumber field is set. */ boolean hasFragmentNumber(); /** * optional int32 fragment_number = 3; + * @return The fragmentNumber. */ int getFragmentNumber(); - // optional int32 attempt_number = 4; /** * optional int32 attempt_number = 4; + * @return Whether the attemptNumber field is set. */ boolean hasAttemptNumber(); /** * optional int32 attempt_number = 4; + * @return The attemptNumber. */ int getAttemptNumber(); - // optional string container_id_string = 5; /** * optional string container_id_string = 5; + * @return Whether the containerIdString field is set. */ boolean hasContainerIdString(); /** * optional string container_id_string = 5; + * @return The containerIdString. */ java.lang.String getContainerIdString(); /** * optional string container_id_string = 5; + * @return The bytes for containerIdString. */ com.google.protobuf.ByteString getContainerIdStringBytes(); - // optional string am_host = 6; /** * optional string am_host = 6; + * @return Whether the amHost field is set. */ boolean hasAmHost(); /** * optional string am_host = 6; + * @return The amHost. */ java.lang.String getAmHost(); /** * optional string am_host = 6; + * @return The bytes for amHost. */ com.google.protobuf.ByteString getAmHostBytes(); - // optional int32 am_port = 7; /** * optional int32 am_port = 7; + * @return Whether the amPort field is set. */ boolean hasAmPort(); /** * optional int32 am_port = 7; + * @return The amPort. */ int getAmPort(); - // optional bytes credentials_binary = 8; /** - * optional bytes credentials_binary = 8; - * *
      * Credentials are not signed - the client can add e.g. his own HDFS tokens.
      * 
+ * + * optional bytes credentials_binary = 8; + * @return Whether the credentialsBinary field is set. */ boolean hasCredentialsBinary(); /** - * optional bytes credentials_binary = 8; - * *
      * Credentials are not signed - the client can add e.g. his own HDFS tokens.
      * 
+ * + * optional bytes credentials_binary = 8; + * @return The credentialsBinary. */ com.google.protobuf.ByteString getCredentialsBinary(); - // optional .FragmentRuntimeInfo fragment_runtime_info = 9; /** - * optional .FragmentRuntimeInfo fragment_runtime_info = 9; - * *
      * Not supported/honored for external clients right now.
      * 
+ * + * optional .FragmentRuntimeInfo fragment_runtime_info = 9; + * @return Whether the fragmentRuntimeInfo field is set. */ boolean hasFragmentRuntimeInfo(); /** - * optional .FragmentRuntimeInfo fragment_runtime_info = 9; - * *
      * Not supported/honored for external clients right now.
      * 
+ * + * optional .FragmentRuntimeInfo fragment_runtime_info = 9; + * @return The fragmentRuntimeInfo. */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo getFragmentRuntimeInfo(); /** - * optional .FragmentRuntimeInfo fragment_runtime_info = 9; - * *
      * Not supported/honored for external clients right now.
      * 
+ * + * optional .FragmentRuntimeInfo fragment_runtime_info = 9; */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfoOrBuilder getFragmentRuntimeInfoOrBuilder(); - // optional bytes initial_event_bytes = 10; /** - * optional bytes initial_event_bytes = 10; - * *
      * Serialized (and signed) NotTezEvent; used only for external clients for now.
      * 
+ * + * optional bytes initial_event_bytes = 10; + * @return Whether the initialEventBytes field is set. */ boolean hasInitialEventBytes(); /** - * optional bytes initial_event_bytes = 10; - * *
      * Serialized (and signed) NotTezEvent; used only for external clients for now.
      * 
+ * + * optional bytes initial_event_bytes = 10; + * @return The initialEventBytes. */ com.google.protobuf.ByteString getInitialEventBytes(); - // optional bytes initial_event_signature = 11; /** * optional bytes initial_event_signature = 11; + * @return Whether the initialEventSignature field is set. */ boolean hasInitialEventSignature(); /** * optional bytes initial_event_signature = 11; + * @return The initialEventSignature. */ com.google.protobuf.ByteString getInitialEventSignature(); - // optional bool is_guaranteed = 12 [default = false]; /** * optional bool is_guaranteed = 12 [default = false]; + * @return Whether the isGuaranteed field is set. */ boolean hasIsGuaranteed(); /** * optional bool is_guaranteed = 12 [default = false]; + * @return The isGuaranteed. */ boolean getIsGuaranteed(); - // optional string jwt = 13; /** * optional string jwt = 13; + * @return Whether the jwt field is set. */ boolean hasJwt(); /** * optional string jwt = 13; + * @return The jwt. */ java.lang.String getJwt(); /** * optional string jwt = 13; + * @return The bytes for jwt. */ com.google.protobuf.ByteString getJwtBytes(); - // optional bool is_external_client_request = 14 [default = false]; /** * optional bool is_external_client_request = 14 [default = false]; + * @return Whether the isExternalClientRequest field is set. */ boolean hasIsExternalClientRequest(); /** * optional bool is_external_client_request = 14 [default = false]; + * @return The isExternalClientRequest. */ boolean getIsExternalClientRequest(); } @@ -9731,35 +10946,44 @@ public interface SubmitWorkRequestProtoOrBuilder * Protobuf type {@code SubmitWorkRequestProto} */ public static final class SubmitWorkRequestProto extends - com.google.protobuf.GeneratedMessage - implements SubmitWorkRequestProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:SubmitWorkRequestProto) + SubmitWorkRequestProtoOrBuilder { + private static final long serialVersionUID = 0L; // Use SubmitWorkRequestProto.newBuilder() to construct. - private SubmitWorkRequestProto(com.google.protobuf.GeneratedMessage.Builder builder) { + private SubmitWorkRequestProto(com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private SubmitWorkRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final SubmitWorkRequestProto defaultInstance; - public static SubmitWorkRequestProto getDefaultInstance() { - return defaultInstance; + private SubmitWorkRequestProto() { + workSpecSignature_ = com.google.protobuf.ByteString.EMPTY; + containerIdString_ = ""; + amHost_ = ""; + credentialsBinary_ = com.google.protobuf.ByteString.EMPTY; + initialEventBytes_ = com.google.protobuf.ByteString.EMPTY; + initialEventSignature_ = com.google.protobuf.ByteString.EMPTY; + jwt_ = ""; } - public SubmitWorkRequestProto getDefaultInstanceForType() { - return defaultInstance; + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance( + UnusedPrivateParameter unused) { + return new SubmitWorkRequestProto(); } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private SubmitWorkRequestProto( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); @@ -9771,16 +10995,9 @@ private SubmitWorkRequestProto( case 0: done = true; break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } case 10: { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary.Builder subBuilder = null; - if (((bitField0_ & 0x00000001) == 0x00000001)) { + if (((bitField0_ & 0x00000001) != 0)) { subBuilder = workSpec_.toBuilder(); } workSpec_ = input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary.PARSER, extensionRegistry); @@ -9807,13 +11024,15 @@ private SubmitWorkRequestProto( break; } case 42: { + com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000010; - containerIdString_ = input.readBytes(); + containerIdString_ = bs; break; } case 50: { + com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000020; - amHost_ = input.readBytes(); + amHost_ = bs; break; } case 56: { @@ -9828,7 +11047,7 @@ private SubmitWorkRequestProto( } case 74: { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo.Builder subBuilder = null; - if (((bitField0_ & 0x00000100) == 0x00000100)) { + if (((bitField0_ & 0x00000100) != 0)) { subBuilder = fragmentRuntimeInfo_.toBuilder(); } fragmentRuntimeInfo_ = input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo.PARSER, extensionRegistry); @@ -9855,8 +11074,9 @@ private SubmitWorkRequestProto( break; } case 106: { + com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00001000; - jwt_ = input.readBytes(); + jwt_ = bs; break; } case 112: { @@ -9864,13 +11084,20 @@ private SubmitWorkRequestProto( isExternalClientRequest_ = input.readBool(); break; } + default: { + if (!parseUnknownField( + input, unknownFields, extensionRegistry, tag)) { + done = true; + } + break; + } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); @@ -9881,111 +11108,113 @@ private SubmitWorkRequestProto( return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SubmitWorkRequestProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SubmitWorkRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public SubmitWorkRequestProto parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new SubmitWorkRequestProto(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - private int bitField0_; - // optional .VertexOrBinary work_spec = 1; public static final int WORK_SPEC_FIELD_NUMBER = 1; private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary workSpec_; /** * optional .VertexOrBinary work_spec = 1; + * @return Whether the workSpec field is set. */ + @java.lang.Override public boolean hasWorkSpec() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional .VertexOrBinary work_spec = 1; + * @return The workSpec. */ + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary getWorkSpec() { - return workSpec_; + return workSpec_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary.getDefaultInstance() : workSpec_; } /** * optional .VertexOrBinary work_spec = 1; */ + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinaryOrBuilder getWorkSpecOrBuilder() { - return workSpec_; + return workSpec_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary.getDefaultInstance() : workSpec_; } - // optional bytes work_spec_signature = 2; public static final int WORK_SPEC_SIGNATURE_FIELD_NUMBER = 2; private com.google.protobuf.ByteString workSpecSignature_; /** * optional bytes work_spec_signature = 2; + * @return Whether the workSpecSignature field is set. */ + @java.lang.Override public boolean hasWorkSpecSignature() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * optional bytes work_spec_signature = 2; + * @return The workSpecSignature. */ + @java.lang.Override public com.google.protobuf.ByteString getWorkSpecSignature() { return workSpecSignature_; } - // optional int32 fragment_number = 3; public static final int FRAGMENT_NUMBER_FIELD_NUMBER = 3; private int fragmentNumber_; /** * optional int32 fragment_number = 3; + * @return Whether the fragmentNumber field is set. */ + @java.lang.Override public boolean hasFragmentNumber() { - return ((bitField0_ & 0x00000004) == 0x00000004); + return ((bitField0_ & 0x00000004) != 0); } /** * optional int32 fragment_number = 3; + * @return The fragmentNumber. */ + @java.lang.Override public int getFragmentNumber() { return fragmentNumber_; } - // optional int32 attempt_number = 4; public static final int ATTEMPT_NUMBER_FIELD_NUMBER = 4; private int attemptNumber_; /** * optional int32 attempt_number = 4; + * @return Whether the attemptNumber field is set. */ + @java.lang.Override public boolean hasAttemptNumber() { - return ((bitField0_ & 0x00000008) == 0x00000008); + return ((bitField0_ & 0x00000008) != 0); } /** * optional int32 attempt_number = 4; + * @return The attemptNumber. */ + @java.lang.Override public int getAttemptNumber() { return attemptNumber_; } - // optional string container_id_string = 5; public static final int CONTAINER_ID_STRING_FIELD_NUMBER = 5; - private java.lang.Object containerIdString_; + private volatile java.lang.Object containerIdString_; /** * optional string container_id_string = 5; + * @return Whether the containerIdString field is set. */ + @java.lang.Override public boolean hasContainerIdString() { - return ((bitField0_ & 0x00000010) == 0x00000010); + return ((bitField0_ & 0x00000010) != 0); } /** * optional string container_id_string = 5; + * @return The containerIdString. */ + @java.lang.Override public java.lang.String getContainerIdString() { java.lang.Object ref = containerIdString_; if (ref instanceof java.lang.String) { @@ -10002,7 +11231,9 @@ public java.lang.String getContainerIdString() { } /** * optional string container_id_string = 5; + * @return The bytes for containerIdString. */ + @java.lang.Override public com.google.protobuf.ByteString getContainerIdStringBytes() { java.lang.Object ref = containerIdString_; @@ -10017,18 +11248,21 @@ public java.lang.String getContainerIdString() { } } - // optional string am_host = 6; public static final int AM_HOST_FIELD_NUMBER = 6; - private java.lang.Object amHost_; + private volatile java.lang.Object amHost_; /** * optional string am_host = 6; + * @return Whether the amHost field is set. */ + @java.lang.Override public boolean hasAmHost() { - return ((bitField0_ & 0x00000020) == 0x00000020); + return ((bitField0_ & 0x00000020) != 0); } /** * optional string am_host = 6; + * @return The amHost. */ + @java.lang.Override public java.lang.String getAmHost() { java.lang.Object ref = amHost_; if (ref instanceof java.lang.String) { @@ -10045,7 +11279,9 @@ public java.lang.String getAmHost() { } /** * optional string am_host = 6; + * @return The bytes for amHost. */ + @java.lang.Override public com.google.protobuf.ByteString getAmHostBytes() { java.lang.Object ref = amHost_; @@ -10060,148 +11296,170 @@ public java.lang.String getAmHost() { } } - // optional int32 am_port = 7; public static final int AM_PORT_FIELD_NUMBER = 7; private int amPort_; /** * optional int32 am_port = 7; + * @return Whether the amPort field is set. */ + @java.lang.Override public boolean hasAmPort() { - return ((bitField0_ & 0x00000040) == 0x00000040); + return ((bitField0_ & 0x00000040) != 0); } /** * optional int32 am_port = 7; + * @return The amPort. */ + @java.lang.Override public int getAmPort() { return amPort_; } - // optional bytes credentials_binary = 8; public static final int CREDENTIALS_BINARY_FIELD_NUMBER = 8; private com.google.protobuf.ByteString credentialsBinary_; /** - * optional bytes credentials_binary = 8; - * *
      * Credentials are not signed - the client can add e.g. his own HDFS tokens.
      * 
+ * + * optional bytes credentials_binary = 8; + * @return Whether the credentialsBinary field is set. */ + @java.lang.Override public boolean hasCredentialsBinary() { - return ((bitField0_ & 0x00000080) == 0x00000080); + return ((bitField0_ & 0x00000080) != 0); } /** - * optional bytes credentials_binary = 8; - * *
      * Credentials are not signed - the client can add e.g. his own HDFS tokens.
      * 
+ * + * optional bytes credentials_binary = 8; + * @return The credentialsBinary. */ + @java.lang.Override public com.google.protobuf.ByteString getCredentialsBinary() { return credentialsBinary_; } - // optional .FragmentRuntimeInfo fragment_runtime_info = 9; public static final int FRAGMENT_RUNTIME_INFO_FIELD_NUMBER = 9; private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo fragmentRuntimeInfo_; /** - * optional .FragmentRuntimeInfo fragment_runtime_info = 9; - * *
      * Not supported/honored for external clients right now.
      * 
+ * + * optional .FragmentRuntimeInfo fragment_runtime_info = 9; + * @return Whether the fragmentRuntimeInfo field is set. */ + @java.lang.Override public boolean hasFragmentRuntimeInfo() { - return ((bitField0_ & 0x00000100) == 0x00000100); + return ((bitField0_ & 0x00000100) != 0); } /** - * optional .FragmentRuntimeInfo fragment_runtime_info = 9; - * *
      * Not supported/honored for external clients right now.
      * 
+ * + * optional .FragmentRuntimeInfo fragment_runtime_info = 9; + * @return The fragmentRuntimeInfo. */ + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo getFragmentRuntimeInfo() { - return fragmentRuntimeInfo_; + return fragmentRuntimeInfo_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo.getDefaultInstance() : fragmentRuntimeInfo_; } /** - * optional .FragmentRuntimeInfo fragment_runtime_info = 9; - * *
      * Not supported/honored for external clients right now.
      * 
+ * + * optional .FragmentRuntimeInfo fragment_runtime_info = 9; */ + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfoOrBuilder getFragmentRuntimeInfoOrBuilder() { - return fragmentRuntimeInfo_; + return fragmentRuntimeInfo_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo.getDefaultInstance() : fragmentRuntimeInfo_; } - // optional bytes initial_event_bytes = 10; public static final int INITIAL_EVENT_BYTES_FIELD_NUMBER = 10; private com.google.protobuf.ByteString initialEventBytes_; /** - * optional bytes initial_event_bytes = 10; - * *
      * Serialized (and signed) NotTezEvent; used only for external clients for now.
      * 
+ * + * optional bytes initial_event_bytes = 10; + * @return Whether the initialEventBytes field is set. */ + @java.lang.Override public boolean hasInitialEventBytes() { - return ((bitField0_ & 0x00000200) == 0x00000200); + return ((bitField0_ & 0x00000200) != 0); } /** - * optional bytes initial_event_bytes = 10; - * *
      * Serialized (and signed) NotTezEvent; used only for external clients for now.
      * 
+ * + * optional bytes initial_event_bytes = 10; + * @return The initialEventBytes. */ + @java.lang.Override public com.google.protobuf.ByteString getInitialEventBytes() { return initialEventBytes_; } - // optional bytes initial_event_signature = 11; public static final int INITIAL_EVENT_SIGNATURE_FIELD_NUMBER = 11; private com.google.protobuf.ByteString initialEventSignature_; /** * optional bytes initial_event_signature = 11; + * @return Whether the initialEventSignature field is set. */ + @java.lang.Override public boolean hasInitialEventSignature() { - return ((bitField0_ & 0x00000400) == 0x00000400); + return ((bitField0_ & 0x00000400) != 0); } /** * optional bytes initial_event_signature = 11; + * @return The initialEventSignature. */ + @java.lang.Override public com.google.protobuf.ByteString getInitialEventSignature() { return initialEventSignature_; } - // optional bool is_guaranteed = 12 [default = false]; public static final int IS_GUARANTEED_FIELD_NUMBER = 12; private boolean isGuaranteed_; /** * optional bool is_guaranteed = 12 [default = false]; + * @return Whether the isGuaranteed field is set. */ + @java.lang.Override public boolean hasIsGuaranteed() { - return ((bitField0_ & 0x00000800) == 0x00000800); + return ((bitField0_ & 0x00000800) != 0); } /** * optional bool is_guaranteed = 12 [default = false]; + * @return The isGuaranteed. */ + @java.lang.Override public boolean getIsGuaranteed() { return isGuaranteed_; } - // optional string jwt = 13; public static final int JWT_FIELD_NUMBER = 13; - private java.lang.Object jwt_; + private volatile java.lang.Object jwt_; /** * optional string jwt = 13; + * @return Whether the jwt field is set. */ + @java.lang.Override public boolean hasJwt() { - return ((bitField0_ & 0x00001000) == 0x00001000); + return ((bitField0_ & 0x00001000) != 0); } /** * optional string jwt = 13; + * @return The jwt. */ + @java.lang.Override public java.lang.String getJwt() { java.lang.Object ref = jwt_; if (ref instanceof java.lang.String) { @@ -10218,7 +11476,9 @@ public java.lang.String getJwt() { } /** * optional string jwt = 13; + * @return The bytes for jwt. */ + @java.lang.Override public com.google.protobuf.ByteString getJwtBytes() { java.lang.Object ref = jwt_; @@ -10233,169 +11493,148 @@ public java.lang.String getJwt() { } } - // optional bool is_external_client_request = 14 [default = false]; public static final int IS_EXTERNAL_CLIENT_REQUEST_FIELD_NUMBER = 14; private boolean isExternalClientRequest_; /** * optional bool is_external_client_request = 14 [default = false]; + * @return Whether the isExternalClientRequest field is set. */ + @java.lang.Override public boolean hasIsExternalClientRequest() { - return ((bitField0_ & 0x00002000) == 0x00002000); + return ((bitField0_ & 0x00002000) != 0); } /** * optional bool is_external_client_request = 14 [default = false]; + * @return The isExternalClientRequest. */ + @java.lang.Override public boolean getIsExternalClientRequest() { return isExternalClientRequest_; } - private void initFields() { - workSpec_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary.getDefaultInstance(); - workSpecSignature_ = com.google.protobuf.ByteString.EMPTY; - fragmentNumber_ = 0; - attemptNumber_ = 0; - containerIdString_ = ""; - amHost_ = ""; - amPort_ = 0; - credentialsBinary_ = com.google.protobuf.ByteString.EMPTY; - fragmentRuntimeInfo_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo.getDefaultInstance(); - initialEventBytes_ = com.google.protobuf.ByteString.EMPTY; - initialEventSignature_ = com.google.protobuf.ByteString.EMPTY; - isGuaranteed_ = false; - jwt_ = ""; - isExternalClientRequest_ = false; - } private byte memoizedIsInitialized = -1; + @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } + @java.lang.Override public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); - if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeMessage(1, workSpec_); + if (((bitField0_ & 0x00000001) != 0)) { + output.writeMessage(1, getWorkSpec()); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { + if (((bitField0_ & 0x00000002) != 0)) { output.writeBytes(2, workSpecSignature_); } - if (((bitField0_ & 0x00000004) == 0x00000004)) { + if (((bitField0_ & 0x00000004) != 0)) { output.writeInt32(3, fragmentNumber_); } - if (((bitField0_ & 0x00000008) == 0x00000008)) { + if (((bitField0_ & 0x00000008) != 0)) { output.writeInt32(4, attemptNumber_); } - if (((bitField0_ & 0x00000010) == 0x00000010)) { - output.writeBytes(5, getContainerIdStringBytes()); + if (((bitField0_ & 0x00000010) != 0)) { + com.google.protobuf.GeneratedMessageV3.writeString(output, 5, containerIdString_); } - if (((bitField0_ & 0x00000020) == 0x00000020)) { - output.writeBytes(6, getAmHostBytes()); + if (((bitField0_ & 0x00000020) != 0)) { + com.google.protobuf.GeneratedMessageV3.writeString(output, 6, amHost_); } - if (((bitField0_ & 0x00000040) == 0x00000040)) { + if (((bitField0_ & 0x00000040) != 0)) { output.writeInt32(7, amPort_); } - if (((bitField0_ & 0x00000080) == 0x00000080)) { + if (((bitField0_ & 0x00000080) != 0)) { output.writeBytes(8, credentialsBinary_); } - if (((bitField0_ & 0x00000100) == 0x00000100)) { - output.writeMessage(9, fragmentRuntimeInfo_); + if (((bitField0_ & 0x00000100) != 0)) { + output.writeMessage(9, getFragmentRuntimeInfo()); } - if (((bitField0_ & 0x00000200) == 0x00000200)) { + if (((bitField0_ & 0x00000200) != 0)) { output.writeBytes(10, initialEventBytes_); } - if (((bitField0_ & 0x00000400) == 0x00000400)) { + if (((bitField0_ & 0x00000400) != 0)) { output.writeBytes(11, initialEventSignature_); } - if (((bitField0_ & 0x00000800) == 0x00000800)) { + if (((bitField0_ & 0x00000800) != 0)) { output.writeBool(12, isGuaranteed_); } - if (((bitField0_ & 0x00001000) == 0x00001000)) { - output.writeBytes(13, getJwtBytes()); + if (((bitField0_ & 0x00001000) != 0)) { + com.google.protobuf.GeneratedMessageV3.writeString(output, 13, jwt_); } - if (((bitField0_ & 0x00002000) == 0x00002000)) { + if (((bitField0_ & 0x00002000) != 0)) { output.writeBool(14, isExternalClientRequest_); } - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; + @java.lang.Override public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { + if (((bitField0_ & 0x00000001) != 0)) { size += com.google.protobuf.CodedOutputStream - .computeMessageSize(1, workSpec_); + .computeMessageSize(1, getWorkSpec()); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { + if (((bitField0_ & 0x00000002) != 0)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(2, workSpecSignature_); } - if (((bitField0_ & 0x00000004) == 0x00000004)) { + if (((bitField0_ & 0x00000004) != 0)) { size += com.google.protobuf.CodedOutputStream .computeInt32Size(3, fragmentNumber_); } - if (((bitField0_ & 0x00000008) == 0x00000008)) { + if (((bitField0_ & 0x00000008) != 0)) { size += com.google.protobuf.CodedOutputStream .computeInt32Size(4, attemptNumber_); } - if (((bitField0_ & 0x00000010) == 0x00000010)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(5, getContainerIdStringBytes()); + if (((bitField0_ & 0x00000010) != 0)) { + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(5, containerIdString_); } - if (((bitField0_ & 0x00000020) == 0x00000020)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(6, getAmHostBytes()); + if (((bitField0_ & 0x00000020) != 0)) { + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(6, amHost_); } - if (((bitField0_ & 0x00000040) == 0x00000040)) { + if (((bitField0_ & 0x00000040) != 0)) { size += com.google.protobuf.CodedOutputStream .computeInt32Size(7, amPort_); } - if (((bitField0_ & 0x00000080) == 0x00000080)) { + if (((bitField0_ & 0x00000080) != 0)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(8, credentialsBinary_); } - if (((bitField0_ & 0x00000100) == 0x00000100)) { + if (((bitField0_ & 0x00000100) != 0)) { size += com.google.protobuf.CodedOutputStream - .computeMessageSize(9, fragmentRuntimeInfo_); + .computeMessageSize(9, getFragmentRuntimeInfo()); } - if (((bitField0_ & 0x00000200) == 0x00000200)) { + if (((bitField0_ & 0x00000200) != 0)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(10, initialEventBytes_); } - if (((bitField0_ & 0x00000400) == 0x00000400)) { + if (((bitField0_ & 0x00000400) != 0)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(11, initialEventSignature_); } - if (((bitField0_ & 0x00000800) == 0x00000800)) { + if (((bitField0_ & 0x00000800) != 0)) { size += com.google.protobuf.CodedOutputStream .computeBoolSize(12, isGuaranteed_); } - if (((bitField0_ & 0x00001000) == 0x00001000)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(13, getJwtBytes()); + if (((bitField0_ & 0x00001000) != 0)) { + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(13, jwt_); } - if (((bitField0_ & 0x00002000) == 0x00002000)) { + if (((bitField0_ & 0x00002000) != 0)) { size += com.google.protobuf.CodedOutputStream .computeBoolSize(14, isExternalClientRequest_); } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { @@ -10406,90 +11645,87 @@ public boolean equals(final java.lang.Object obj) { } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto) obj; - boolean result = true; - result = result && (hasWorkSpec() == other.hasWorkSpec()); + if (hasWorkSpec() != other.hasWorkSpec()) return false; if (hasWorkSpec()) { - result = result && getWorkSpec() - .equals(other.getWorkSpec()); + if (!getWorkSpec() + .equals(other.getWorkSpec())) return false; } - result = result && (hasWorkSpecSignature() == other.hasWorkSpecSignature()); + if (hasWorkSpecSignature() != other.hasWorkSpecSignature()) return false; if (hasWorkSpecSignature()) { - result = result && getWorkSpecSignature() - .equals(other.getWorkSpecSignature()); + if (!getWorkSpecSignature() + .equals(other.getWorkSpecSignature())) return false; } - result = result && (hasFragmentNumber() == other.hasFragmentNumber()); + if (hasFragmentNumber() != other.hasFragmentNumber()) return false; if (hasFragmentNumber()) { - result = result && (getFragmentNumber() - == other.getFragmentNumber()); + if (getFragmentNumber() + != other.getFragmentNumber()) return false; } - result = result && (hasAttemptNumber() == other.hasAttemptNumber()); + if (hasAttemptNumber() != other.hasAttemptNumber()) return false; if (hasAttemptNumber()) { - result = result && (getAttemptNumber() - == other.getAttemptNumber()); + if (getAttemptNumber() + != other.getAttemptNumber()) return false; } - result = result && (hasContainerIdString() == other.hasContainerIdString()); + if (hasContainerIdString() != other.hasContainerIdString()) return false; if (hasContainerIdString()) { - result = result && getContainerIdString() - .equals(other.getContainerIdString()); + if (!getContainerIdString() + .equals(other.getContainerIdString())) return false; } - result = result && (hasAmHost() == other.hasAmHost()); + if (hasAmHost() != other.hasAmHost()) return false; if (hasAmHost()) { - result = result && getAmHost() - .equals(other.getAmHost()); + if (!getAmHost() + .equals(other.getAmHost())) return false; } - result = result && (hasAmPort() == other.hasAmPort()); + if (hasAmPort() != other.hasAmPort()) return false; if (hasAmPort()) { - result = result && (getAmPort() - == other.getAmPort()); + if (getAmPort() + != other.getAmPort()) return false; } - result = result && (hasCredentialsBinary() == other.hasCredentialsBinary()); + if (hasCredentialsBinary() != other.hasCredentialsBinary()) return false; if (hasCredentialsBinary()) { - result = result && getCredentialsBinary() - .equals(other.getCredentialsBinary()); + if (!getCredentialsBinary() + .equals(other.getCredentialsBinary())) return false; } - result = result && (hasFragmentRuntimeInfo() == other.hasFragmentRuntimeInfo()); + if (hasFragmentRuntimeInfo() != other.hasFragmentRuntimeInfo()) return false; if (hasFragmentRuntimeInfo()) { - result = result && getFragmentRuntimeInfo() - .equals(other.getFragmentRuntimeInfo()); + if (!getFragmentRuntimeInfo() + .equals(other.getFragmentRuntimeInfo())) return false; } - result = result && (hasInitialEventBytes() == other.hasInitialEventBytes()); + if (hasInitialEventBytes() != other.hasInitialEventBytes()) return false; if (hasInitialEventBytes()) { - result = result && getInitialEventBytes() - .equals(other.getInitialEventBytes()); + if (!getInitialEventBytes() + .equals(other.getInitialEventBytes())) return false; } - result = result && (hasInitialEventSignature() == other.hasInitialEventSignature()); + if (hasInitialEventSignature() != other.hasInitialEventSignature()) return false; if (hasInitialEventSignature()) { - result = result && getInitialEventSignature() - .equals(other.getInitialEventSignature()); + if (!getInitialEventSignature() + .equals(other.getInitialEventSignature())) return false; } - result = result && (hasIsGuaranteed() == other.hasIsGuaranteed()); + if (hasIsGuaranteed() != other.hasIsGuaranteed()) return false; if (hasIsGuaranteed()) { - result = result && (getIsGuaranteed() - == other.getIsGuaranteed()); + if (getIsGuaranteed() + != other.getIsGuaranteed()) return false; } - result = result && (hasJwt() == other.hasJwt()); + if (hasJwt() != other.hasJwt()) return false; if (hasJwt()) { - result = result && getJwt() - .equals(other.getJwt()); + if (!getJwt() + .equals(other.getJwt())) return false; } - result = result && (hasIsExternalClientRequest() == other.hasIsExternalClientRequest()); + if (hasIsExternalClientRequest() != other.hasIsExternalClientRequest()) return false; if (hasIsExternalClientRequest()) { - result = result && (getIsExternalClientRequest() - == other.getIsExternalClientRequest()); + if (getIsExternalClientRequest() + != other.getIsExternalClientRequest()) return false; } - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; + if (!unknownFields.equals(other.unknownFields)) return false; + return true; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); + hash = (19 * hash) + getDescriptor().hashCode(); if (hasWorkSpec()) { hash = (37 * hash) + WORK_SPEC_FIELD_NUMBER; hash = (53 * hash) + getWorkSpec().hashCode(); @@ -10536,7 +11772,8 @@ public int hashCode() { } if (hasIsGuaranteed()) { hash = (37 * hash) + IS_GUARANTEED_FIELD_NUMBER; - hash = (53 * hash) + hashBoolean(getIsGuaranteed()); + hash = (53 * hash) + com.google.protobuf.Internal.hashBoolean( + getIsGuaranteed()); } if (hasJwt()) { hash = (37 * hash) + JWT_FIELD_NUMBER; @@ -10544,13 +11781,25 @@ public int hashCode() { } if (hasIsExternalClientRequest()) { hash = (37 * hash) + IS_EXTERNAL_CLIENT_REQUEST_FIELD_NUMBER; - hash = (53 * hash) + hashBoolean(getIsExternalClientRequest()); + hash = (53 * hash) + com.google.protobuf.Internal.hashBoolean( + getIsExternalClientRequest()); } - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto parseFrom( + java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto parseFrom( + java.nio.ByteBuffer data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { @@ -10574,46 +11823,59 @@ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.Su } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } + @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -10621,14 +11883,16 @@ protected Builder newBuilderForType( * Protobuf type {@code SubmitWorkRequestProto} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder implements + // @@protoc_insertion_point(builder_implements:SubmitWorkRequestProto) + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProtoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SubmitWorkRequestProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SubmitWorkRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -10641,24 +11905,22 @@ private Builder() { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { getWorkSpecFieldBuilder(); getFragmentRuntimeInfoFieldBuilder(); } } - private static Builder create() { - return new Builder(); - } - + @java.lang.Override public Builder clear() { super.clear(); if (workSpecBuilder_ == null) { - workSpec_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary.getDefaultInstance(); + workSpec_ = null; } else { workSpecBuilder_.clear(); } @@ -10678,7 +11940,7 @@ public Builder clear() { credentialsBinary_ = com.google.protobuf.ByteString.EMPTY; bitField0_ = (bitField0_ & ~0x00000080); if (fragmentRuntimeInfoBuilder_ == null) { - fragmentRuntimeInfo_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo.getDefaultInstance(); + fragmentRuntimeInfo_ = null; } else { fragmentRuntimeInfoBuilder_.clear(); } @@ -10696,19 +11958,18 @@ public Builder clear() { return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - + @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SubmitWorkRequestProto_descriptor; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto.getDefaultInstance(); } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto result = buildPartial(); if (!result.isInitialized()) { @@ -10717,79 +11978,113 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWor return result; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { + if (((from_bitField0_ & 0x00000001) != 0)) { + if (workSpecBuilder_ == null) { + result.workSpec_ = workSpec_; + } else { + result.workSpec_ = workSpecBuilder_.build(); + } to_bitField0_ |= 0x00000001; } - if (workSpecBuilder_ == null) { - result.workSpec_ = workSpec_; - } else { - result.workSpec_ = workSpecBuilder_.build(); - } - if (((from_bitField0_ & 0x00000002) == 0x00000002)) { + if (((from_bitField0_ & 0x00000002) != 0)) { to_bitField0_ |= 0x00000002; } result.workSpecSignature_ = workSpecSignature_; - if (((from_bitField0_ & 0x00000004) == 0x00000004)) { + if (((from_bitField0_ & 0x00000004) != 0)) { + result.fragmentNumber_ = fragmentNumber_; to_bitField0_ |= 0x00000004; } - result.fragmentNumber_ = fragmentNumber_; - if (((from_bitField0_ & 0x00000008) == 0x00000008)) { + if (((from_bitField0_ & 0x00000008) != 0)) { + result.attemptNumber_ = attemptNumber_; to_bitField0_ |= 0x00000008; } - result.attemptNumber_ = attemptNumber_; - if (((from_bitField0_ & 0x00000010) == 0x00000010)) { + if (((from_bitField0_ & 0x00000010) != 0)) { to_bitField0_ |= 0x00000010; } result.containerIdString_ = containerIdString_; - if (((from_bitField0_ & 0x00000020) == 0x00000020)) { + if (((from_bitField0_ & 0x00000020) != 0)) { to_bitField0_ |= 0x00000020; } result.amHost_ = amHost_; - if (((from_bitField0_ & 0x00000040) == 0x00000040)) { + if (((from_bitField0_ & 0x00000040) != 0)) { + result.amPort_ = amPort_; to_bitField0_ |= 0x00000040; } - result.amPort_ = amPort_; - if (((from_bitField0_ & 0x00000080) == 0x00000080)) { + if (((from_bitField0_ & 0x00000080) != 0)) { to_bitField0_ |= 0x00000080; } result.credentialsBinary_ = credentialsBinary_; - if (((from_bitField0_ & 0x00000100) == 0x00000100)) { + if (((from_bitField0_ & 0x00000100) != 0)) { + if (fragmentRuntimeInfoBuilder_ == null) { + result.fragmentRuntimeInfo_ = fragmentRuntimeInfo_; + } else { + result.fragmentRuntimeInfo_ = fragmentRuntimeInfoBuilder_.build(); + } to_bitField0_ |= 0x00000100; } - if (fragmentRuntimeInfoBuilder_ == null) { - result.fragmentRuntimeInfo_ = fragmentRuntimeInfo_; - } else { - result.fragmentRuntimeInfo_ = fragmentRuntimeInfoBuilder_.build(); - } - if (((from_bitField0_ & 0x00000200) == 0x00000200)) { + if (((from_bitField0_ & 0x00000200) != 0)) { to_bitField0_ |= 0x00000200; } result.initialEventBytes_ = initialEventBytes_; - if (((from_bitField0_ & 0x00000400) == 0x00000400)) { + if (((from_bitField0_ & 0x00000400) != 0)) { to_bitField0_ |= 0x00000400; } result.initialEventSignature_ = initialEventSignature_; - if (((from_bitField0_ & 0x00000800) == 0x00000800)) { + if (((from_bitField0_ & 0x00000800) != 0)) { + result.isGuaranteed_ = isGuaranteed_; to_bitField0_ |= 0x00000800; } - result.isGuaranteed_ = isGuaranteed_; - if (((from_bitField0_ & 0x00001000) == 0x00001000)) { + if (((from_bitField0_ & 0x00001000) != 0)) { to_bitField0_ |= 0x00001000; } result.jwt_ = jwt_; - if (((from_bitField0_ & 0x00002000) == 0x00002000)) { + if (((from_bitField0_ & 0x00002000) != 0)) { + result.isExternalClientRequest_ = isExternalClientRequest_; to_bitField0_ |= 0x00002000; } - result.isExternalClientRequest_ = isExternalClientRequest_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } + @java.lang.Override + public Builder clone() { + return super.clone(); + } + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.setField(field, value); + } + @java.lang.Override + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + @java.lang.Override + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.addRepeatedField(field, value); + } + @java.lang.Override public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto)other); @@ -10849,14 +12144,17 @@ public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtoc if (other.hasIsExternalClientRequest()) { setIsExternalClientRequest(other.getIsExternalClientRequest()); } - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } + @java.lang.Override public final boolean isInitialized() { return true; } + @java.lang.Override public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) @@ -10866,7 +12164,7 @@ public Builder mergeFrom( parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -10876,22 +12174,23 @@ public Builder mergeFrom( } private int bitField0_; - // optional .VertexOrBinary work_spec = 1; - private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary workSpec_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary.getDefaultInstance(); - private com.google.protobuf.SingleFieldBuilder< + private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary workSpec_; + private com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinaryOrBuilder> workSpecBuilder_; /** * optional .VertexOrBinary work_spec = 1; + * @return Whether the workSpec field is set. */ public boolean hasWorkSpec() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional .VertexOrBinary work_spec = 1; + * @return The workSpec. */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary getWorkSpec() { if (workSpecBuilder_ == null) { - return workSpec_; + return workSpec_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary.getDefaultInstance() : workSpec_; } else { return workSpecBuilder_.getMessage(); } @@ -10931,7 +12230,8 @@ public Builder setWorkSpec( */ public Builder mergeWorkSpec(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary value) { if (workSpecBuilder_ == null) { - if (((bitField0_ & 0x00000001) == 0x00000001) && + if (((bitField0_ & 0x00000001) != 0) && + workSpec_ != null && workSpec_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary.getDefaultInstance()) { workSpec_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary.newBuilder(workSpec_).mergeFrom(value).buildPartial(); @@ -10950,7 +12250,7 @@ public Builder mergeWorkSpec(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonPr */ public Builder clearWorkSpec() { if (workSpecBuilder_ == null) { - workSpec_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary.getDefaultInstance(); + workSpec_ = null; onChanged(); } else { workSpecBuilder_.clear(); @@ -10973,19 +12273,20 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrB if (workSpecBuilder_ != null) { return workSpecBuilder_.getMessageOrBuilder(); } else { - return workSpec_; + return workSpec_ == null ? + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary.getDefaultInstance() : workSpec_; } } /** * optional .VertexOrBinary work_spec = 1; */ - private com.google.protobuf.SingleFieldBuilder< + private com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinaryOrBuilder> getWorkSpecFieldBuilder() { if (workSpecBuilder_ == null) { - workSpecBuilder_ = new com.google.protobuf.SingleFieldBuilder< + workSpecBuilder_ = new com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinary.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrBinaryOrBuilder>( - workSpec_, + getWorkSpec(), getParentForChildren(), isClean()); workSpec_ = null; @@ -10993,22 +12294,27 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.VertexOrB return workSpecBuilder_; } - // optional bytes work_spec_signature = 2; private com.google.protobuf.ByteString workSpecSignature_ = com.google.protobuf.ByteString.EMPTY; /** * optional bytes work_spec_signature = 2; + * @return Whether the workSpecSignature field is set. */ + @java.lang.Override public boolean hasWorkSpecSignature() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * optional bytes work_spec_signature = 2; + * @return The workSpecSignature. */ + @java.lang.Override public com.google.protobuf.ByteString getWorkSpecSignature() { return workSpecSignature_; } /** * optional bytes work_spec_signature = 2; + * @param value The workSpecSignature to set. + * @return This builder for chaining. */ public Builder setWorkSpecSignature(com.google.protobuf.ByteString value) { if (value == null) { @@ -11021,6 +12327,7 @@ public Builder setWorkSpecSignature(com.google.protobuf.ByteString value) { } /** * optional bytes work_spec_signature = 2; + * @return This builder for chaining. */ public Builder clearWorkSpecSignature() { bitField0_ = (bitField0_ & ~0x00000002); @@ -11029,22 +12336,27 @@ public Builder clearWorkSpecSignature() { return this; } - // optional int32 fragment_number = 3; private int fragmentNumber_ ; /** * optional int32 fragment_number = 3; + * @return Whether the fragmentNumber field is set. */ + @java.lang.Override public boolean hasFragmentNumber() { - return ((bitField0_ & 0x00000004) == 0x00000004); + return ((bitField0_ & 0x00000004) != 0); } /** * optional int32 fragment_number = 3; + * @return The fragmentNumber. */ + @java.lang.Override public int getFragmentNumber() { return fragmentNumber_; } /** * optional int32 fragment_number = 3; + * @param value The fragmentNumber to set. + * @return This builder for chaining. */ public Builder setFragmentNumber(int value) { bitField0_ |= 0x00000004; @@ -11054,6 +12366,7 @@ public Builder setFragmentNumber(int value) { } /** * optional int32 fragment_number = 3; + * @return This builder for chaining. */ public Builder clearFragmentNumber() { bitField0_ = (bitField0_ & ~0x00000004); @@ -11062,22 +12375,27 @@ public Builder clearFragmentNumber() { return this; } - // optional int32 attempt_number = 4; private int attemptNumber_ ; /** * optional int32 attempt_number = 4; + * @return Whether the attemptNumber field is set. */ + @java.lang.Override public boolean hasAttemptNumber() { - return ((bitField0_ & 0x00000008) == 0x00000008); + return ((bitField0_ & 0x00000008) != 0); } /** * optional int32 attempt_number = 4; + * @return The attemptNumber. */ + @java.lang.Override public int getAttemptNumber() { return attemptNumber_; } /** * optional int32 attempt_number = 4; + * @param value The attemptNumber to set. + * @return This builder for chaining. */ public Builder setAttemptNumber(int value) { bitField0_ |= 0x00000008; @@ -11087,6 +12405,7 @@ public Builder setAttemptNumber(int value) { } /** * optional int32 attempt_number = 4; + * @return This builder for chaining. */ public Builder clearAttemptNumber() { bitField0_ = (bitField0_ & ~0x00000008); @@ -11095,23 +12414,27 @@ public Builder clearAttemptNumber() { return this; } - // optional string container_id_string = 5; private java.lang.Object containerIdString_ = ""; /** * optional string container_id_string = 5; + * @return Whether the containerIdString field is set. */ public boolean hasContainerIdString() { - return ((bitField0_ & 0x00000010) == 0x00000010); + return ((bitField0_ & 0x00000010) != 0); } /** * optional string container_id_string = 5; + * @return The containerIdString. */ public java.lang.String getContainerIdString() { java.lang.Object ref = containerIdString_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - containerIdString_ = s; + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + containerIdString_ = s; + } return s; } else { return (java.lang.String) ref; @@ -11119,6 +12442,7 @@ public java.lang.String getContainerIdString() { } /** * optional string container_id_string = 5; + * @return The bytes for containerIdString. */ public com.google.protobuf.ByteString getContainerIdStringBytes() { @@ -11135,6 +12459,8 @@ public java.lang.String getContainerIdString() { } /** * optional string container_id_string = 5; + * @param value The containerIdString to set. + * @return This builder for chaining. */ public Builder setContainerIdString( java.lang.String value) { @@ -11148,6 +12474,7 @@ public Builder setContainerIdString( } /** * optional string container_id_string = 5; + * @return This builder for chaining. */ public Builder clearContainerIdString() { bitField0_ = (bitField0_ & ~0x00000010); @@ -11157,6 +12484,8 @@ public Builder clearContainerIdString() { } /** * optional string container_id_string = 5; + * @param value The bytes for containerIdString to set. + * @return This builder for chaining. */ public Builder setContainerIdStringBytes( com.google.protobuf.ByteString value) { @@ -11169,23 +12498,27 @@ public Builder setContainerIdStringBytes( return this; } - // optional string am_host = 6; private java.lang.Object amHost_ = ""; /** * optional string am_host = 6; + * @return Whether the amHost field is set. */ public boolean hasAmHost() { - return ((bitField0_ & 0x00000020) == 0x00000020); + return ((bitField0_ & 0x00000020) != 0); } /** * optional string am_host = 6; + * @return The amHost. */ public java.lang.String getAmHost() { java.lang.Object ref = amHost_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - amHost_ = s; + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + amHost_ = s; + } return s; } else { return (java.lang.String) ref; @@ -11193,6 +12526,7 @@ public java.lang.String getAmHost() { } /** * optional string am_host = 6; + * @return The bytes for amHost. */ public com.google.protobuf.ByteString getAmHostBytes() { @@ -11209,6 +12543,8 @@ public java.lang.String getAmHost() { } /** * optional string am_host = 6; + * @param value The amHost to set. + * @return This builder for chaining. */ public Builder setAmHost( java.lang.String value) { @@ -11222,6 +12558,7 @@ public Builder setAmHost( } /** * optional string am_host = 6; + * @return This builder for chaining. */ public Builder clearAmHost() { bitField0_ = (bitField0_ & ~0x00000020); @@ -11231,6 +12568,8 @@ public Builder clearAmHost() { } /** * optional string am_host = 6; + * @param value The bytes for amHost to set. + * @return This builder for chaining. */ public Builder setAmHostBytes( com.google.protobuf.ByteString value) { @@ -11243,22 +12582,27 @@ public Builder setAmHostBytes( return this; } - // optional int32 am_port = 7; private int amPort_ ; /** * optional int32 am_port = 7; + * @return Whether the amPort field is set. */ + @java.lang.Override public boolean hasAmPort() { - return ((bitField0_ & 0x00000040) == 0x00000040); + return ((bitField0_ & 0x00000040) != 0); } /** * optional int32 am_port = 7; + * @return The amPort. */ + @java.lang.Override public int getAmPort() { return amPort_; } /** * optional int32 am_port = 7; + * @param value The amPort to set. + * @return This builder for chaining. */ public Builder setAmPort(int value) { bitField0_ |= 0x00000040; @@ -11268,6 +12612,7 @@ public Builder setAmPort(int value) { } /** * optional int32 am_port = 7; + * @return This builder for chaining. */ public Builder clearAmPort() { bitField0_ = (bitField0_ & ~0x00000040); @@ -11276,34 +12621,39 @@ public Builder clearAmPort() { return this; } - // optional bytes credentials_binary = 8; private com.google.protobuf.ByteString credentialsBinary_ = com.google.protobuf.ByteString.EMPTY; /** - * optional bytes credentials_binary = 8; - * *
        * Credentials are not signed - the client can add e.g. his own HDFS tokens.
        * 
+ * + * optional bytes credentials_binary = 8; + * @return Whether the credentialsBinary field is set. */ + @java.lang.Override public boolean hasCredentialsBinary() { - return ((bitField0_ & 0x00000080) == 0x00000080); + return ((bitField0_ & 0x00000080) != 0); } /** - * optional bytes credentials_binary = 8; - * *
        * Credentials are not signed - the client can add e.g. his own HDFS tokens.
        * 
+ * + * optional bytes credentials_binary = 8; + * @return The credentialsBinary. */ + @java.lang.Override public com.google.protobuf.ByteString getCredentialsBinary() { return credentialsBinary_; } /** - * optional bytes credentials_binary = 8; - * *
        * Credentials are not signed - the client can add e.g. his own HDFS tokens.
        * 
+ * + * optional bytes credentials_binary = 8; + * @param value The credentialsBinary to set. + * @return This builder for chaining. */ public Builder setCredentialsBinary(com.google.protobuf.ByteString value) { if (value == null) { @@ -11315,11 +12665,12 @@ public Builder setCredentialsBinary(com.google.protobuf.ByteString value) { return this; } /** - * optional bytes credentials_binary = 8; - * *
        * Credentials are not signed - the client can add e.g. his own HDFS tokens.
        * 
+ * + * optional bytes credentials_binary = 8; + * @return This builder for chaining. */ public Builder clearCredentialsBinary() { bitField0_ = (bitField0_ & ~0x00000080); @@ -11328,40 +12679,41 @@ public Builder clearCredentialsBinary() { return this; } - // optional .FragmentRuntimeInfo fragment_runtime_info = 9; - private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo fragmentRuntimeInfo_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo.getDefaultInstance(); - private com.google.protobuf.SingleFieldBuilder< + private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo fragmentRuntimeInfo_; + private com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfoOrBuilder> fragmentRuntimeInfoBuilder_; /** - * optional .FragmentRuntimeInfo fragment_runtime_info = 9; - * *
        * Not supported/honored for external clients right now.
        * 
+ * + * optional .FragmentRuntimeInfo fragment_runtime_info = 9; + * @return Whether the fragmentRuntimeInfo field is set. */ public boolean hasFragmentRuntimeInfo() { - return ((bitField0_ & 0x00000100) == 0x00000100); + return ((bitField0_ & 0x00000100) != 0); } /** - * optional .FragmentRuntimeInfo fragment_runtime_info = 9; - * *
        * Not supported/honored for external clients right now.
        * 
+ * + * optional .FragmentRuntimeInfo fragment_runtime_info = 9; + * @return The fragmentRuntimeInfo. */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo getFragmentRuntimeInfo() { if (fragmentRuntimeInfoBuilder_ == null) { - return fragmentRuntimeInfo_; + return fragmentRuntimeInfo_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo.getDefaultInstance() : fragmentRuntimeInfo_; } else { return fragmentRuntimeInfoBuilder_.getMessage(); } } /** - * optional .FragmentRuntimeInfo fragment_runtime_info = 9; - * *
        * Not supported/honored for external clients right now.
        * 
+ * + * optional .FragmentRuntimeInfo fragment_runtime_info = 9; */ public Builder setFragmentRuntimeInfo(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo value) { if (fragmentRuntimeInfoBuilder_ == null) { @@ -11377,11 +12729,11 @@ public Builder setFragmentRuntimeInfo(org.apache.hadoop.hive.llap.daemon.rpc.Lla return this; } /** - * optional .FragmentRuntimeInfo fragment_runtime_info = 9; - * *
        * Not supported/honored for external clients right now.
        * 
+ * + * optional .FragmentRuntimeInfo fragment_runtime_info = 9; */ public Builder setFragmentRuntimeInfo( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo.Builder builderForValue) { @@ -11395,15 +12747,16 @@ public Builder setFragmentRuntimeInfo( return this; } /** - * optional .FragmentRuntimeInfo fragment_runtime_info = 9; - * *
        * Not supported/honored for external clients right now.
        * 
+ * + * optional .FragmentRuntimeInfo fragment_runtime_info = 9; */ public Builder mergeFragmentRuntimeInfo(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo value) { if (fragmentRuntimeInfoBuilder_ == null) { - if (((bitField0_ & 0x00000100) == 0x00000100) && + if (((bitField0_ & 0x00000100) != 0) && + fragmentRuntimeInfo_ != null && fragmentRuntimeInfo_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo.getDefaultInstance()) { fragmentRuntimeInfo_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo.newBuilder(fragmentRuntimeInfo_).mergeFrom(value).buildPartial(); @@ -11418,15 +12771,15 @@ public Builder mergeFragmentRuntimeInfo(org.apache.hadoop.hive.llap.daemon.rpc.L return this; } /** - * optional .FragmentRuntimeInfo fragment_runtime_info = 9; - * *
        * Not supported/honored for external clients right now.
        * 
+ * + * optional .FragmentRuntimeInfo fragment_runtime_info = 9; */ public Builder clearFragmentRuntimeInfo() { if (fragmentRuntimeInfoBuilder_ == null) { - fragmentRuntimeInfo_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo.getDefaultInstance(); + fragmentRuntimeInfo_ = null; onChanged(); } else { fragmentRuntimeInfoBuilder_.clear(); @@ -11435,11 +12788,11 @@ public Builder clearFragmentRuntimeInfo() { return this; } /** - * optional .FragmentRuntimeInfo fragment_runtime_info = 9; - * *
        * Not supported/honored for external clients right now.
        * 
+ * + * optional .FragmentRuntimeInfo fragment_runtime_info = 9; */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo.Builder getFragmentRuntimeInfoBuilder() { bitField0_ |= 0x00000100; @@ -11447,33 +12800,34 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentR return getFragmentRuntimeInfoFieldBuilder().getBuilder(); } /** - * optional .FragmentRuntimeInfo fragment_runtime_info = 9; - * *
        * Not supported/honored for external clients right now.
        * 
+ * + * optional .FragmentRuntimeInfo fragment_runtime_info = 9; */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfoOrBuilder getFragmentRuntimeInfoOrBuilder() { if (fragmentRuntimeInfoBuilder_ != null) { return fragmentRuntimeInfoBuilder_.getMessageOrBuilder(); } else { - return fragmentRuntimeInfo_; + return fragmentRuntimeInfo_ == null ? + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo.getDefaultInstance() : fragmentRuntimeInfo_; } } /** - * optional .FragmentRuntimeInfo fragment_runtime_info = 9; - * *
        * Not supported/honored for external clients right now.
        * 
+ * + * optional .FragmentRuntimeInfo fragment_runtime_info = 9; */ - private com.google.protobuf.SingleFieldBuilder< + private com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfoOrBuilder> getFragmentRuntimeInfoFieldBuilder() { if (fragmentRuntimeInfoBuilder_ == null) { - fragmentRuntimeInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder< + fragmentRuntimeInfoBuilder_ = new com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfo.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentRuntimeInfoOrBuilder>( - fragmentRuntimeInfo_, + getFragmentRuntimeInfo(), getParentForChildren(), isClean()); fragmentRuntimeInfo_ = null; @@ -11481,34 +12835,39 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.FragmentR return fragmentRuntimeInfoBuilder_; } - // optional bytes initial_event_bytes = 10; private com.google.protobuf.ByteString initialEventBytes_ = com.google.protobuf.ByteString.EMPTY; /** - * optional bytes initial_event_bytes = 10; - * *
        * Serialized (and signed) NotTezEvent; used only for external clients for now.
        * 
+ * + * optional bytes initial_event_bytes = 10; + * @return Whether the initialEventBytes field is set. */ + @java.lang.Override public boolean hasInitialEventBytes() { - return ((bitField0_ & 0x00000200) == 0x00000200); + return ((bitField0_ & 0x00000200) != 0); } /** - * optional bytes initial_event_bytes = 10; - * *
        * Serialized (and signed) NotTezEvent; used only for external clients for now.
        * 
+ * + * optional bytes initial_event_bytes = 10; + * @return The initialEventBytes. */ + @java.lang.Override public com.google.protobuf.ByteString getInitialEventBytes() { return initialEventBytes_; } /** - * optional bytes initial_event_bytes = 10; - * *
        * Serialized (and signed) NotTezEvent; used only for external clients for now.
        * 
+ * + * optional bytes initial_event_bytes = 10; + * @param value The initialEventBytes to set. + * @return This builder for chaining. */ public Builder setInitialEventBytes(com.google.protobuf.ByteString value) { if (value == null) { @@ -11520,11 +12879,12 @@ public Builder setInitialEventBytes(com.google.protobuf.ByteString value) { return this; } /** - * optional bytes initial_event_bytes = 10; - * *
        * Serialized (and signed) NotTezEvent; used only for external clients for now.
        * 
+ * + * optional bytes initial_event_bytes = 10; + * @return This builder for chaining. */ public Builder clearInitialEventBytes() { bitField0_ = (bitField0_ & ~0x00000200); @@ -11533,22 +12893,27 @@ public Builder clearInitialEventBytes() { return this; } - // optional bytes initial_event_signature = 11; private com.google.protobuf.ByteString initialEventSignature_ = com.google.protobuf.ByteString.EMPTY; /** * optional bytes initial_event_signature = 11; + * @return Whether the initialEventSignature field is set. */ + @java.lang.Override public boolean hasInitialEventSignature() { - return ((bitField0_ & 0x00000400) == 0x00000400); + return ((bitField0_ & 0x00000400) != 0); } /** * optional bytes initial_event_signature = 11; + * @return The initialEventSignature. */ + @java.lang.Override public com.google.protobuf.ByteString getInitialEventSignature() { return initialEventSignature_; } /** * optional bytes initial_event_signature = 11; + * @param value The initialEventSignature to set. + * @return This builder for chaining. */ public Builder setInitialEventSignature(com.google.protobuf.ByteString value) { if (value == null) { @@ -11561,6 +12926,7 @@ public Builder setInitialEventSignature(com.google.protobuf.ByteString value) { } /** * optional bytes initial_event_signature = 11; + * @return This builder for chaining. */ public Builder clearInitialEventSignature() { bitField0_ = (bitField0_ & ~0x00000400); @@ -11569,22 +12935,27 @@ public Builder clearInitialEventSignature() { return this; } - // optional bool is_guaranteed = 12 [default = false]; private boolean isGuaranteed_ ; /** * optional bool is_guaranteed = 12 [default = false]; + * @return Whether the isGuaranteed field is set. */ + @java.lang.Override public boolean hasIsGuaranteed() { - return ((bitField0_ & 0x00000800) == 0x00000800); + return ((bitField0_ & 0x00000800) != 0); } /** * optional bool is_guaranteed = 12 [default = false]; + * @return The isGuaranteed. */ + @java.lang.Override public boolean getIsGuaranteed() { return isGuaranteed_; } /** * optional bool is_guaranteed = 12 [default = false]; + * @param value The isGuaranteed to set. + * @return This builder for chaining. */ public Builder setIsGuaranteed(boolean value) { bitField0_ |= 0x00000800; @@ -11594,6 +12965,7 @@ public Builder setIsGuaranteed(boolean value) { } /** * optional bool is_guaranteed = 12 [default = false]; + * @return This builder for chaining. */ public Builder clearIsGuaranteed() { bitField0_ = (bitField0_ & ~0x00000800); @@ -11602,23 +12974,27 @@ public Builder clearIsGuaranteed() { return this; } - // optional string jwt = 13; private java.lang.Object jwt_ = ""; /** * optional string jwt = 13; + * @return Whether the jwt field is set. */ public boolean hasJwt() { - return ((bitField0_ & 0x00001000) == 0x00001000); + return ((bitField0_ & 0x00001000) != 0); } /** * optional string jwt = 13; + * @return The jwt. */ public java.lang.String getJwt() { java.lang.Object ref = jwt_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - jwt_ = s; + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + jwt_ = s; + } return s; } else { return (java.lang.String) ref; @@ -11626,6 +13002,7 @@ public java.lang.String getJwt() { } /** * optional string jwt = 13; + * @return The bytes for jwt. */ public com.google.protobuf.ByteString getJwtBytes() { @@ -11642,6 +13019,8 @@ public java.lang.String getJwt() { } /** * optional string jwt = 13; + * @param value The jwt to set. + * @return This builder for chaining. */ public Builder setJwt( java.lang.String value) { @@ -11655,6 +13034,7 @@ public Builder setJwt( } /** * optional string jwt = 13; + * @return This builder for chaining. */ public Builder clearJwt() { bitField0_ = (bitField0_ & ~0x00001000); @@ -11664,6 +13044,8 @@ public Builder clearJwt() { } /** * optional string jwt = 13; + * @param value The bytes for jwt to set. + * @return This builder for chaining. */ public Builder setJwtBytes( com.google.protobuf.ByteString value) { @@ -11676,22 +13058,27 @@ public Builder setJwtBytes( return this; } - // optional bool is_external_client_request = 14 [default = false]; private boolean isExternalClientRequest_ ; /** * optional bool is_external_client_request = 14 [default = false]; + * @return Whether the isExternalClientRequest field is set. */ + @java.lang.Override public boolean hasIsExternalClientRequest() { - return ((bitField0_ & 0x00002000) == 0x00002000); + return ((bitField0_ & 0x00002000) != 0); } /** * optional bool is_external_client_request = 14 [default = false]; + * @return The isExternalClientRequest. */ + @java.lang.Override public boolean getIsExternalClientRequest() { return isExternalClientRequest_; } /** * optional bool is_external_client_request = 14 [default = false]; + * @param value The isExternalClientRequest to set. + * @return This builder for chaining. */ public Builder setIsExternalClientRequest(boolean value) { bitField0_ |= 0x00002000; @@ -11701,6 +13088,7 @@ public Builder setIsExternalClientRequest(boolean value) { } /** * optional bool is_external_client_request = 14 [default = false]; + * @return This builder for chaining. */ public Builder clearIsExternalClientRequest() { bitField0_ = (bitField0_ & ~0x00002000); @@ -11708,43 +13096,88 @@ public Builder clearIsExternalClientRequest() { onChanged(); return this; } + @java.lang.Override + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:SubmitWorkRequestProto) } + // @@protoc_insertion_point(class_scope:SubmitWorkRequestProto) + private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto DEFAULT_INSTANCE; static { - defaultInstance = new SubmitWorkRequestProto(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto(); + } + + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser + PARSER = new com.google.protobuf.AbstractParser() { + @java.lang.Override + public SubmitWorkRequestProto parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new SubmitWorkRequestProto(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkRequestProto getDefaultInstanceForType() { + return DEFAULT_INSTANCE; } - // @@protoc_insertion_point(class_scope:SubmitWorkRequestProto) } - public interface RegisterDagRequestProtoOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface RegisterDagRequestProtoOrBuilder extends + // @@protoc_insertion_point(interface_extends:RegisterDagRequestProto) + com.google.protobuf.MessageOrBuilder { - // optional string user = 1; /** * optional string user = 1; + * @return Whether the user field is set. */ boolean hasUser(); /** * optional string user = 1; + * @return The user. */ java.lang.String getUser(); /** * optional string user = 1; + * @return The bytes for user. */ com.google.protobuf.ByteString getUserBytes(); - // required .QueryIdentifierProto query_identifier = 2; /** * required .QueryIdentifierProto query_identifier = 2; + * @return Whether the queryIdentifier field is set. */ boolean hasQueryIdentifier(); /** * required .QueryIdentifierProto query_identifier = 2; + * @return The queryIdentifier. */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto getQueryIdentifier(); /** @@ -11752,13 +13185,14 @@ public interface RegisterDagRequestProtoOrBuilder */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder getQueryIdentifierOrBuilder(); - // optional bytes credentials_binary = 3; /** * optional bytes credentials_binary = 3; + * @return Whether the credentialsBinary field is set. */ boolean hasCredentialsBinary(); /** * optional bytes credentials_binary = 3; + * @return The credentialsBinary. */ com.google.protobuf.ByteString getCredentialsBinary(); } @@ -11766,35 +13200,39 @@ public interface RegisterDagRequestProtoOrBuilder * Protobuf type {@code RegisterDagRequestProto} */ public static final class RegisterDagRequestProto extends - com.google.protobuf.GeneratedMessage - implements RegisterDagRequestProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:RegisterDagRequestProto) + RegisterDagRequestProtoOrBuilder { + private static final long serialVersionUID = 0L; // Use RegisterDagRequestProto.newBuilder() to construct. - private RegisterDagRequestProto(com.google.protobuf.GeneratedMessage.Builder builder) { + private RegisterDagRequestProto(com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private RegisterDagRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final RegisterDagRequestProto defaultInstance; - public static RegisterDagRequestProto getDefaultInstance() { - return defaultInstance; + private RegisterDagRequestProto() { + user_ = ""; + credentialsBinary_ = com.google.protobuf.ByteString.EMPTY; } - public RegisterDagRequestProto getDefaultInstanceForType() { - return defaultInstance; + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance( + UnusedPrivateParameter unused) { + return new RegisterDagRequestProto(); } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private RegisterDagRequestProto( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); @@ -11806,21 +13244,15 @@ private RegisterDagRequestProto( case 0: done = true; break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } case 10: { + com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000001; - user_ = input.readBytes(); + user_ = bs; break; } case 18: { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder subBuilder = null; - if (((bitField0_ & 0x00000002) == 0x00000002)) { + if (((bitField0_ & 0x00000002) != 0)) { subBuilder = queryIdentifier_.toBuilder(); } queryIdentifier_ = input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.PARSER, extensionRegistry); @@ -11836,13 +13268,20 @@ private RegisterDagRequestProto( credentialsBinary_ = input.readBytes(); break; } + default: { + if (!parseUnknownField( + input, unknownFields, extensionRegistry, tag)) { + done = true; + } + break; + } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); @@ -11853,41 +13292,30 @@ private RegisterDagRequestProto( return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_RegisterDagRequestProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_RegisterDagRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagRequestProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagRequestProto.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public RegisterDagRequestProto parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new RegisterDagRequestProto(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - private int bitField0_; - // optional string user = 1; public static final int USER_FIELD_NUMBER = 1; - private java.lang.Object user_; + private volatile java.lang.Object user_; /** * optional string user = 1; + * @return Whether the user field is set. */ + @java.lang.Override public boolean hasUser() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional string user = 1; + * @return The user. */ + @java.lang.Override public java.lang.String getUser() { java.lang.Object ref = user_; if (ref instanceof java.lang.String) { @@ -11904,7 +13332,9 @@ public java.lang.String getUser() { } /** * optional string user = 1; + * @return The bytes for user. */ + @java.lang.Override public com.google.protobuf.ByteString getUserBytes() { java.lang.Object ref = user_; @@ -11919,53 +13349,57 @@ public java.lang.String getUser() { } } - // required .QueryIdentifierProto query_identifier = 2; public static final int QUERY_IDENTIFIER_FIELD_NUMBER = 2; private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto queryIdentifier_; /** * required .QueryIdentifierProto query_identifier = 2; + * @return Whether the queryIdentifier field is set. */ + @java.lang.Override public boolean hasQueryIdentifier() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * required .QueryIdentifierProto query_identifier = 2; + * @return The queryIdentifier. */ + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto getQueryIdentifier() { - return queryIdentifier_; + return queryIdentifier_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance() : queryIdentifier_; } /** * required .QueryIdentifierProto query_identifier = 2; */ + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder getQueryIdentifierOrBuilder() { - return queryIdentifier_; + return queryIdentifier_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance() : queryIdentifier_; } - // optional bytes credentials_binary = 3; public static final int CREDENTIALS_BINARY_FIELD_NUMBER = 3; private com.google.protobuf.ByteString credentialsBinary_; /** * optional bytes credentials_binary = 3; + * @return Whether the credentialsBinary field is set. */ + @java.lang.Override public boolean hasCredentialsBinary() { - return ((bitField0_ & 0x00000004) == 0x00000004); + return ((bitField0_ & 0x00000004) != 0); } /** * optional bytes credentials_binary = 3; + * @return The credentialsBinary. */ + @java.lang.Override public com.google.protobuf.ByteString getCredentialsBinary() { return credentialsBinary_; } - private void initFields() { - user_ = ""; - queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance(); - credentialsBinary_ = com.google.protobuf.ByteString.EMPTY; - } private byte memoizedIsInitialized = -1; + @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; if (!hasQueryIdentifier()) { memoizedIsInitialized = 0; @@ -11975,51 +13409,43 @@ public final boolean isInitialized() { return true; } + @java.lang.Override public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); - if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeBytes(1, getUserBytes()); + if (((bitField0_ & 0x00000001) != 0)) { + com.google.protobuf.GeneratedMessageV3.writeString(output, 1, user_); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - output.writeMessage(2, queryIdentifier_); + if (((bitField0_ & 0x00000002) != 0)) { + output.writeMessage(2, getQueryIdentifier()); } - if (((bitField0_ & 0x00000004) == 0x00000004)) { + if (((bitField0_ & 0x00000004) != 0)) { output.writeBytes(3, credentialsBinary_); } - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; + @java.lang.Override public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(1, getUserBytes()); + if (((bitField0_ & 0x00000001) != 0)) { + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, user_); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { + if (((bitField0_ & 0x00000002) != 0)) { size += com.google.protobuf.CodedOutputStream - .computeMessageSize(2, queryIdentifier_); + .computeMessageSize(2, getQueryIdentifier()); } - if (((bitField0_ & 0x00000004) == 0x00000004)) { + if (((bitField0_ & 0x00000004) != 0)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(3, credentialsBinary_); } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { @@ -12030,35 +13456,32 @@ public boolean equals(final java.lang.Object obj) { } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagRequestProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagRequestProto) obj; - boolean result = true; - result = result && (hasUser() == other.hasUser()); + if (hasUser() != other.hasUser()) return false; if (hasUser()) { - result = result && getUser() - .equals(other.getUser()); + if (!getUser() + .equals(other.getUser())) return false; } - result = result && (hasQueryIdentifier() == other.hasQueryIdentifier()); + if (hasQueryIdentifier() != other.hasQueryIdentifier()) return false; if (hasQueryIdentifier()) { - result = result && getQueryIdentifier() - .equals(other.getQueryIdentifier()); + if (!getQueryIdentifier() + .equals(other.getQueryIdentifier())) return false; } - result = result && (hasCredentialsBinary() == other.hasCredentialsBinary()); + if (hasCredentialsBinary() != other.hasCredentialsBinary()) return false; if (hasCredentialsBinary()) { - result = result && getCredentialsBinary() - .equals(other.getCredentialsBinary()); + if (!getCredentialsBinary() + .equals(other.getCredentialsBinary())) return false; } - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; + if (!unknownFields.equals(other.unknownFields)) return false; + return true; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); + hash = (19 * hash) + getDescriptor().hashCode(); if (hasUser()) { hash = (37 * hash) + USER_FIELD_NUMBER; hash = (53 * hash) + getUser().hashCode(); @@ -12071,11 +13494,22 @@ public int hashCode() { hash = (37 * hash) + CREDENTIALS_BINARY_FIELD_NUMBER; hash = (53 * hash) + getCredentialsBinary().hashCode(); } - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagRequestProto parseFrom( + java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagRequestProto parseFrom( + java.nio.ByteBuffer data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagRequestProto parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { @@ -12099,46 +13533,59 @@ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.Re } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagRequestProto parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagRequestProto parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagRequestProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagRequestProto parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagRequestProto parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagRequestProto parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } + @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagRequestProto prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -12146,14 +13593,16 @@ protected Builder newBuilderForType( * Protobuf type {@code RegisterDagRequestProto} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagRequestProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder implements + // @@protoc_insertion_point(builder_implements:RegisterDagRequestProto) + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagRequestProtoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_RegisterDagRequestProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_RegisterDagRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -12166,25 +13615,23 @@ private Builder() { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { getQueryIdentifierFieldBuilder(); } } - private static Builder create() { - return new Builder(); - } - + @java.lang.Override public Builder clear() { super.clear(); user_ = ""; bitField0_ = (bitField0_ & ~0x00000001); if (queryIdentifierBuilder_ == null) { - queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance(); + queryIdentifier_ = null; } else { queryIdentifierBuilder_.clear(); } @@ -12194,19 +13641,18 @@ public Builder clear() { return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - + @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_RegisterDagRequestProto_descriptor; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagRequestProto getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagRequestProto.getDefaultInstance(); } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagRequestProto build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagRequestProto result = buildPartial(); if (!result.isInitialized()) { @@ -12215,23 +13661,24 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterD return result; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagRequestProto buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagRequestProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagRequestProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { + if (((from_bitField0_ & 0x00000001) != 0)) { to_bitField0_ |= 0x00000001; } result.user_ = user_; - if (((from_bitField0_ & 0x00000002) == 0x00000002)) { + if (((from_bitField0_ & 0x00000002) != 0)) { + if (queryIdentifierBuilder_ == null) { + result.queryIdentifier_ = queryIdentifier_; + } else { + result.queryIdentifier_ = queryIdentifierBuilder_.build(); + } to_bitField0_ |= 0x00000002; } - if (queryIdentifierBuilder_ == null) { - result.queryIdentifier_ = queryIdentifier_; - } else { - result.queryIdentifier_ = queryIdentifierBuilder_.build(); - } - if (((from_bitField0_ & 0x00000004) == 0x00000004)) { + if (((from_bitField0_ & 0x00000004) != 0)) { to_bitField0_ |= 0x00000004; } result.credentialsBinary_ = credentialsBinary_; @@ -12240,6 +13687,39 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterD return result; } + @java.lang.Override + public Builder clone() { + return super.clone(); + } + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.setField(field, value); + } + @java.lang.Override + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + @java.lang.Override + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.addRepeatedField(field, value); + } + @java.lang.Override public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagRequestProto) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagRequestProto)other); @@ -12262,18 +13742,20 @@ public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtoc if (other.hasCredentialsBinary()) { setCredentialsBinary(other.getCredentialsBinary()); } - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } + @java.lang.Override public final boolean isInitialized() { if (!hasQueryIdentifier()) { - return false; } return true; } + @java.lang.Override public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) @@ -12283,7 +13765,7 @@ public Builder mergeFrom( parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagRequestProto) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -12293,23 +13775,27 @@ public Builder mergeFrom( } private int bitField0_; - // optional string user = 1; private java.lang.Object user_ = ""; /** * optional string user = 1; + * @return Whether the user field is set. */ public boolean hasUser() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional string user = 1; + * @return The user. */ public java.lang.String getUser() { java.lang.Object ref = user_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - user_ = s; + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + user_ = s; + } return s; } else { return (java.lang.String) ref; @@ -12317,6 +13803,7 @@ public java.lang.String getUser() { } /** * optional string user = 1; + * @return The bytes for user. */ public com.google.protobuf.ByteString getUserBytes() { @@ -12333,6 +13820,8 @@ public java.lang.String getUser() { } /** * optional string user = 1; + * @param value The user to set. + * @return This builder for chaining. */ public Builder setUser( java.lang.String value) { @@ -12346,6 +13835,7 @@ public Builder setUser( } /** * optional string user = 1; + * @return This builder for chaining. */ public Builder clearUser() { bitField0_ = (bitField0_ & ~0x00000001); @@ -12355,6 +13845,8 @@ public Builder clearUser() { } /** * optional string user = 1; + * @param value The bytes for user to set. + * @return This builder for chaining. */ public Builder setUserBytes( com.google.protobuf.ByteString value) { @@ -12367,22 +13859,23 @@ public Builder setUserBytes( return this; } - // required .QueryIdentifierProto query_identifier = 2; - private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance(); - private com.google.protobuf.SingleFieldBuilder< + private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto queryIdentifier_; + private com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder> queryIdentifierBuilder_; /** * required .QueryIdentifierProto query_identifier = 2; + * @return Whether the queryIdentifier field is set. */ public boolean hasQueryIdentifier() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * required .QueryIdentifierProto query_identifier = 2; + * @return The queryIdentifier. */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto getQueryIdentifier() { if (queryIdentifierBuilder_ == null) { - return queryIdentifier_; + return queryIdentifier_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance() : queryIdentifier_; } else { return queryIdentifierBuilder_.getMessage(); } @@ -12422,7 +13915,8 @@ public Builder setQueryIdentifier( */ public Builder mergeQueryIdentifier(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto value) { if (queryIdentifierBuilder_ == null) { - if (((bitField0_ & 0x00000002) == 0x00000002) && + if (((bitField0_ & 0x00000002) != 0) && + queryIdentifier_ != null && queryIdentifier_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance()) { queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.newBuilder(queryIdentifier_).mergeFrom(value).buildPartial(); @@ -12441,7 +13935,7 @@ public Builder mergeQueryIdentifier(org.apache.hadoop.hive.llap.daemon.rpc.LlapD */ public Builder clearQueryIdentifier() { if (queryIdentifierBuilder_ == null) { - queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance(); + queryIdentifier_ = null; onChanged(); } else { queryIdentifierBuilder_.clear(); @@ -12464,19 +13958,20 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIden if (queryIdentifierBuilder_ != null) { return queryIdentifierBuilder_.getMessageOrBuilder(); } else { - return queryIdentifier_; + return queryIdentifier_ == null ? + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance() : queryIdentifier_; } } /** * required .QueryIdentifierProto query_identifier = 2; */ - private com.google.protobuf.SingleFieldBuilder< + private com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder> getQueryIdentifierFieldBuilder() { if (queryIdentifierBuilder_ == null) { - queryIdentifierBuilder_ = new com.google.protobuf.SingleFieldBuilder< + queryIdentifierBuilder_ = new com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder>( - queryIdentifier_, + getQueryIdentifier(), getParentForChildren(), isClean()); queryIdentifier_ = null; @@ -12484,22 +13979,27 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIden return queryIdentifierBuilder_; } - // optional bytes credentials_binary = 3; private com.google.protobuf.ByteString credentialsBinary_ = com.google.protobuf.ByteString.EMPTY; /** * optional bytes credentials_binary = 3; + * @return Whether the credentialsBinary field is set. */ + @java.lang.Override public boolean hasCredentialsBinary() { - return ((bitField0_ & 0x00000004) == 0x00000004); + return ((bitField0_ & 0x00000004) != 0); } /** * optional bytes credentials_binary = 3; + * @return The credentialsBinary. */ + @java.lang.Override public com.google.protobuf.ByteString getCredentialsBinary() { return credentialsBinary_; } /** * optional bytes credentials_binary = 3; + * @param value The credentialsBinary to set. + * @return This builder for chaining. */ public Builder setCredentialsBinary(com.google.protobuf.ByteString value) { if (value == null) { @@ -12512,6 +14012,7 @@ public Builder setCredentialsBinary(com.google.protobuf.ByteString value) { } /** * optional bytes credentials_binary = 3; + * @return This builder for chaining. */ public Builder clearCredentialsBinary() { bitField0_ = (bitField0_ & ~0x00000004); @@ -12519,54 +14020,98 @@ public Builder clearCredentialsBinary() { onChanged(); return this; } + @java.lang.Override + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } - // @@protoc_insertion_point(builder_scope:RegisterDagRequestProto) - } - - static { - defaultInstance = new RegisterDagRequestProto(true); - defaultInstance.initFields(); - } + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } - // @@protoc_insertion_point(class_scope:RegisterDagRequestProto) - } - public interface RegisterDagResponseProtoOrBuilder - extends com.google.protobuf.MessageOrBuilder { - } - /** - * Protobuf type {@code RegisterDagResponseProto} - */ - public static final class RegisterDagResponseProto extends - com.google.protobuf.GeneratedMessage - implements RegisterDagResponseProtoOrBuilder { - // Use RegisterDagResponseProto.newBuilder() to construct. - private RegisterDagResponseProto(com.google.protobuf.GeneratedMessage.Builder builder) { - super(builder); - this.unknownFields = builder.getUnknownFields(); + // @@protoc_insertion_point(builder_scope:RegisterDagRequestProto) } - private RegisterDagResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - private static final RegisterDagResponseProto defaultInstance; - public static RegisterDagResponseProto getDefaultInstance() { - return defaultInstance; + // @@protoc_insertion_point(class_scope:RegisterDagRequestProto) + private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagRequestProto DEFAULT_INSTANCE; + static { + DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagRequestProto(); } - public RegisterDagResponseProto getDefaultInstanceForType() { - return defaultInstance; + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagRequestProto getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser + PARSER = new com.google.protobuf.AbstractParser() { + @java.lang.Override + public RegisterDagRequestProto parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new RegisterDagRequestProto(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagRequestProto getDefaultInstanceForType() { + return DEFAULT_INSTANCE; + } + + } + + public interface RegisterDagResponseProtoOrBuilder extends + // @@protoc_insertion_point(interface_extends:RegisterDagResponseProto) + com.google.protobuf.MessageOrBuilder { + } + /** + * Protobuf type {@code RegisterDagResponseProto} + */ + public static final class RegisterDagResponseProto extends + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:RegisterDagResponseProto) + RegisterDagResponseProtoOrBuilder { + private static final long serialVersionUID = 0L; + // Use RegisterDagResponseProto.newBuilder() to construct. + private RegisterDagResponseProto(com.google.protobuf.GeneratedMessageV3.Builder builder) { + super(builder); + } + private RegisterDagResponseProto() { + } + + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance( + UnusedPrivateParameter unused) { + return new RegisterDagResponseProto(); } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private RegisterDagResponseProto( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); try { @@ -12578,8 +14123,8 @@ private RegisterDagResponseProto( done = true; break; default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { + if (!parseUnknownField( + input, unknownFields, extensionRegistry, tag)) { done = true; } break; @@ -12590,7 +14135,7 @@ private RegisterDagResponseProto( throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); @@ -12601,63 +14146,42 @@ private RegisterDagResponseProto( return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_RegisterDagResponseProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_RegisterDagResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagResponseProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagResponseProto.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public RegisterDagResponseProto parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new RegisterDagResponseProto(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - - private void initFields() { - } private byte memoizedIsInitialized = -1; + @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } + @java.lang.Override public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; + @java.lang.Override public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { @@ -12668,25 +14192,33 @@ public boolean equals(final java.lang.Object obj) { } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagResponseProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagResponseProto) obj; - boolean result = true; - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; + if (!unknownFields.equals(other.unknownFields)) return false; + return true; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (19 * hash) + getDescriptor().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagResponseProto parseFrom( + java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagResponseProto parseFrom( + java.nio.ByteBuffer data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagResponseProto parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { @@ -12710,46 +14242,59 @@ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.Re } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagResponseProto parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagResponseProto parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagResponseProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagResponseProto parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagResponseProto parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagResponseProto parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } + @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagResponseProto prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -12757,14 +14302,16 @@ protected Builder newBuilderForType( * Protobuf type {@code RegisterDagResponseProto} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagResponseProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder implements + // @@protoc_insertion_point(builder_implements:RegisterDagResponseProto) + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagResponseProtoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_RegisterDagResponseProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_RegisterDagResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -12777,36 +14324,33 @@ private Builder() { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { } } - private static Builder create() { - return new Builder(); - } - + @java.lang.Override public Builder clear() { super.clear(); return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - + @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_RegisterDagResponseProto_descriptor; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagResponseProto getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagResponseProto.getDefaultInstance(); } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagResponseProto build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagResponseProto result = buildPartial(); if (!result.isInitialized()) { @@ -12815,12 +14359,46 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterD return result; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagResponseProto buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagResponseProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagResponseProto(this); onBuilt(); return result; } + @java.lang.Override + public Builder clone() { + return super.clone(); + } + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.setField(field, value); + } + @java.lang.Override + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + @java.lang.Override + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.addRepeatedField(field, value); + } + @java.lang.Override public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagResponseProto) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagResponseProto)other); @@ -12832,14 +14410,17 @@ public Builder mergeFrom(com.google.protobuf.Message other) { public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagResponseProto other) { if (other == org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagResponseProto.getDefaultInstance()) return this; - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } + @java.lang.Override public final boolean isInitialized() { return true; } + @java.lang.Override public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) @@ -12849,7 +14430,7 @@ public Builder mergeFrom( parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagResponseProto) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -12857,42 +14438,87 @@ public Builder mergeFrom( } return this; } + @java.lang.Override + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:RegisterDagResponseProto) } + // @@protoc_insertion_point(class_scope:RegisterDagResponseProto) + private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagResponseProto DEFAULT_INSTANCE; static { - defaultInstance = new RegisterDagResponseProto(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagResponseProto(); + } + + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagResponseProto getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser + PARSER = new com.google.protobuf.AbstractParser() { + @java.lang.Override + public RegisterDagResponseProto parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new RegisterDagResponseProto(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.RegisterDagResponseProto getDefaultInstanceForType() { + return DEFAULT_INSTANCE; } - // @@protoc_insertion_point(class_scope:RegisterDagResponseProto) } - public interface SubmitWorkResponseProtoOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface SubmitWorkResponseProtoOrBuilder extends + // @@protoc_insertion_point(interface_extends:SubmitWorkResponseProto) + com.google.protobuf.MessageOrBuilder { - // optional .SubmissionStateProto submission_state = 1; /** * optional .SubmissionStateProto submission_state = 1; + * @return Whether the submissionState field is set. */ boolean hasSubmissionState(); /** * optional .SubmissionStateProto submission_state = 1; + * @return The submissionState. */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmissionStateProto getSubmissionState(); - // optional string unique_node_id = 2; /** * optional string unique_node_id = 2; + * @return Whether the uniqueNodeId field is set. */ boolean hasUniqueNodeId(); /** * optional string unique_node_id = 2; + * @return The uniqueNodeId. */ java.lang.String getUniqueNodeId(); /** * optional string unique_node_id = 2; + * @return The bytes for uniqueNodeId. */ com.google.protobuf.ByteString getUniqueNodeIdBytes(); @@ -12901,35 +14527,39 @@ public interface SubmitWorkResponseProtoOrBuilder * Protobuf type {@code SubmitWorkResponseProto} */ public static final class SubmitWorkResponseProto extends - com.google.protobuf.GeneratedMessage - implements SubmitWorkResponseProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:SubmitWorkResponseProto) + SubmitWorkResponseProtoOrBuilder { + private static final long serialVersionUID = 0L; // Use SubmitWorkResponseProto.newBuilder() to construct. - private SubmitWorkResponseProto(com.google.protobuf.GeneratedMessage.Builder builder) { + private SubmitWorkResponseProto(com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private SubmitWorkResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final SubmitWorkResponseProto defaultInstance; - public static SubmitWorkResponseProto getDefaultInstance() { - return defaultInstance; + private SubmitWorkResponseProto() { + submissionState_ = 1; + uniqueNodeId_ = ""; } - public SubmitWorkResponseProto getDefaultInstanceForType() { - return defaultInstance; + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance( + UnusedPrivateParameter unused) { + return new SubmitWorkResponseProto(); } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private SubmitWorkResponseProto( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); @@ -12941,27 +14571,29 @@ private SubmitWorkResponseProto( case 0: done = true; break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } case 8: { int rawValue = input.readEnum(); + @SuppressWarnings("deprecation") org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmissionStateProto value = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmissionStateProto.valueOf(rawValue); if (value == null) { unknownFields.mergeVarintField(1, rawValue); } else { bitField0_ |= 0x00000001; - submissionState_ = value; + submissionState_ = rawValue; } break; } case 18: { + com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000002; - uniqueNodeId_ = input.readBytes(); + uniqueNodeId_ = bs; + break; + } + default: { + if (!parseUnknownField( + input, unknownFields, extensionRegistry, tag)) { + done = true; + } break; } } @@ -12970,7 +14602,7 @@ private SubmitWorkResponseProto( throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); @@ -12981,57 +14613,49 @@ private SubmitWorkResponseProto( return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SubmitWorkResponseProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SubmitWorkResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public SubmitWorkResponseProto parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new SubmitWorkResponseProto(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - private int bitField0_; - // optional .SubmissionStateProto submission_state = 1; public static final int SUBMISSION_STATE_FIELD_NUMBER = 1; - private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmissionStateProto submissionState_; + private int submissionState_; /** * optional .SubmissionStateProto submission_state = 1; + * @return Whether the submissionState field is set. */ - public boolean hasSubmissionState() { - return ((bitField0_ & 0x00000001) == 0x00000001); + @java.lang.Override public boolean hasSubmissionState() { + return ((bitField0_ & 0x00000001) != 0); } /** * optional .SubmissionStateProto submission_state = 1; + * @return The submissionState. */ - public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmissionStateProto getSubmissionState() { - return submissionState_; + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmissionStateProto getSubmissionState() { + @SuppressWarnings("deprecation") + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmissionStateProto result = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmissionStateProto.valueOf(submissionState_); + return result == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmissionStateProto.ACCEPTED : result; } - // optional string unique_node_id = 2; public static final int UNIQUE_NODE_ID_FIELD_NUMBER = 2; - private java.lang.Object uniqueNodeId_; + private volatile java.lang.Object uniqueNodeId_; /** * optional string unique_node_id = 2; + * @return Whether the uniqueNodeId field is set. */ + @java.lang.Override public boolean hasUniqueNodeId() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * optional string unique_node_id = 2; + * @return The uniqueNodeId. */ + @java.lang.Override public java.lang.String getUniqueNodeId() { java.lang.Object ref = uniqueNodeId_; if (ref instanceof java.lang.String) { @@ -13048,7 +14672,9 @@ public java.lang.String getUniqueNodeId() { } /** * optional string unique_node_id = 2; + * @return The bytes for uniqueNodeId. */ + @java.lang.Override public com.google.protobuf.ByteString getUniqueNodeIdBytes() { java.lang.Object ref = uniqueNodeId_; @@ -13063,57 +14689,47 @@ public java.lang.String getUniqueNodeId() { } } - private void initFields() { - submissionState_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmissionStateProto.ACCEPTED; - uniqueNodeId_ = ""; - } private byte memoizedIsInitialized = -1; + @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } + @java.lang.Override public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); - if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeEnum(1, submissionState_.getNumber()); + if (((bitField0_ & 0x00000001) != 0)) { + output.writeEnum(1, submissionState_); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - output.writeBytes(2, getUniqueNodeIdBytes()); + if (((bitField0_ & 0x00000002) != 0)) { + com.google.protobuf.GeneratedMessageV3.writeString(output, 2, uniqueNodeId_); } - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; + @java.lang.Override public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { + if (((bitField0_ & 0x00000001) != 0)) { size += com.google.protobuf.CodedOutputStream - .computeEnumSize(1, submissionState_.getNumber()); + .computeEnumSize(1, submissionState_); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(2, getUniqueNodeIdBytes()); + if (((bitField0_ & 0x00000002) != 0)) { + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(2, uniqueNodeId_); } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { @@ -13124,43 +14740,50 @@ public boolean equals(final java.lang.Object obj) { } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto) obj; - boolean result = true; - result = result && (hasSubmissionState() == other.hasSubmissionState()); + if (hasSubmissionState() != other.hasSubmissionState()) return false; if (hasSubmissionState()) { - result = result && - (getSubmissionState() == other.getSubmissionState()); + if (submissionState_ != other.submissionState_) return false; } - result = result && (hasUniqueNodeId() == other.hasUniqueNodeId()); + if (hasUniqueNodeId() != other.hasUniqueNodeId()) return false; if (hasUniqueNodeId()) { - result = result && getUniqueNodeId() - .equals(other.getUniqueNodeId()); + if (!getUniqueNodeId() + .equals(other.getUniqueNodeId())) return false; } - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; + if (!unknownFields.equals(other.unknownFields)) return false; + return true; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); + hash = (19 * hash) + getDescriptor().hashCode(); if (hasSubmissionState()) { hash = (37 * hash) + SUBMISSION_STATE_FIELD_NUMBER; - hash = (53 * hash) + hashEnum(getSubmissionState()); + hash = (53 * hash) + submissionState_; } if (hasUniqueNodeId()) { hash = (37 * hash) + UNIQUE_NODE_ID_FIELD_NUMBER; hash = (53 * hash) + getUniqueNodeId().hashCode(); } - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto parseFrom( + java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto parseFrom( + java.nio.ByteBuffer data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { @@ -13184,46 +14807,59 @@ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.Su } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } + @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -13231,14 +14867,16 @@ protected Builder newBuilderForType( * Protobuf type {@code SubmitWorkResponseProto} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder implements + // @@protoc_insertion_point(builder_implements:SubmitWorkResponseProto) + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProtoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SubmitWorkResponseProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SubmitWorkResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -13251,40 +14889,37 @@ private Builder() { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { } } - private static Builder create() { - return new Builder(); - } - + @java.lang.Override public Builder clear() { super.clear(); - submissionState_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmissionStateProto.ACCEPTED; + submissionState_ = 1; bitField0_ = (bitField0_ & ~0x00000001); uniqueNodeId_ = ""; bitField0_ = (bitField0_ & ~0x00000002); return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - + @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SubmitWorkResponseProto_descriptor; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto.getDefaultInstance(); } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto result = buildPartial(); if (!result.isInitialized()) { @@ -13293,15 +14928,16 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWor return result; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { + if (((from_bitField0_ & 0x00000001) != 0)) { to_bitField0_ |= 0x00000001; } result.submissionState_ = submissionState_; - if (((from_bitField0_ & 0x00000002) == 0x00000002)) { + if (((from_bitField0_ & 0x00000002) != 0)) { to_bitField0_ |= 0x00000002; } result.uniqueNodeId_ = uniqueNodeId_; @@ -13310,6 +14946,39 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWor return result; } + @java.lang.Override + public Builder clone() { + return super.clone(); + } + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.setField(field, value); + } + @java.lang.Override + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + @java.lang.Override + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.addRepeatedField(field, value); + } + @java.lang.Override public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto)other); @@ -13329,14 +14998,17 @@ public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtoc uniqueNodeId_ = other.uniqueNodeId_; onChanged(); } - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } + @java.lang.Override public final boolean isInitialized() { return true; } + @java.lang.Override public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) @@ -13346,7 +15018,7 @@ public Builder mergeFrom( parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -13356,59 +15028,70 @@ public Builder mergeFrom( } private int bitField0_; - // optional .SubmissionStateProto submission_state = 1; - private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmissionStateProto submissionState_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmissionStateProto.ACCEPTED; + private int submissionState_ = 1; /** * optional .SubmissionStateProto submission_state = 1; + * @return Whether the submissionState field is set. */ - public boolean hasSubmissionState() { - return ((bitField0_ & 0x00000001) == 0x00000001); + @java.lang.Override public boolean hasSubmissionState() { + return ((bitField0_ & 0x00000001) != 0); } /** * optional .SubmissionStateProto submission_state = 1; + * @return The submissionState. */ + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmissionStateProto getSubmissionState() { - return submissionState_; + @SuppressWarnings("deprecation") + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmissionStateProto result = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmissionStateProto.valueOf(submissionState_); + return result == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmissionStateProto.ACCEPTED : result; } /** * optional .SubmissionStateProto submission_state = 1; + * @param value The submissionState to set. + * @return This builder for chaining. */ public Builder setSubmissionState(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmissionStateProto value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; - submissionState_ = value; + submissionState_ = value.getNumber(); onChanged(); return this; } /** * optional .SubmissionStateProto submission_state = 1; + * @return This builder for chaining. */ public Builder clearSubmissionState() { bitField0_ = (bitField0_ & ~0x00000001); - submissionState_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmissionStateProto.ACCEPTED; + submissionState_ = 1; onChanged(); return this; } - // optional string unique_node_id = 2; private java.lang.Object uniqueNodeId_ = ""; /** * optional string unique_node_id = 2; + * @return Whether the uniqueNodeId field is set. */ public boolean hasUniqueNodeId() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * optional string unique_node_id = 2; + * @return The uniqueNodeId. */ public java.lang.String getUniqueNodeId() { java.lang.Object ref = uniqueNodeId_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - uniqueNodeId_ = s; + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + uniqueNodeId_ = s; + } return s; } else { return (java.lang.String) ref; @@ -13416,6 +15099,7 @@ public java.lang.String getUniqueNodeId() { } /** * optional string unique_node_id = 2; + * @return The bytes for uniqueNodeId. */ public com.google.protobuf.ByteString getUniqueNodeIdBytes() { @@ -13432,6 +15116,8 @@ public java.lang.String getUniqueNodeId() { } /** * optional string unique_node_id = 2; + * @param value The uniqueNodeId to set. + * @return This builder for chaining. */ public Builder setUniqueNodeId( java.lang.String value) { @@ -13445,6 +15131,7 @@ public Builder setUniqueNodeId( } /** * optional string unique_node_id = 2; + * @return This builder for chaining. */ public Builder clearUniqueNodeId() { bitField0_ = (bitField0_ & ~0x00000002); @@ -13454,6 +15141,8 @@ public Builder clearUniqueNodeId() { } /** * optional string unique_node_id = 2; + * @param value The bytes for uniqueNodeId to set. + * @return This builder for chaining. */ public Builder setUniqueNodeIdBytes( com.google.protobuf.ByteString value) { @@ -13465,28 +15154,71 @@ public Builder setUniqueNodeIdBytes( onChanged(); return this; } + @java.lang.Override + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:SubmitWorkResponseProto) } + // @@protoc_insertion_point(class_scope:SubmitWorkResponseProto) + private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto DEFAULT_INSTANCE; static { - defaultInstance = new SubmitWorkResponseProto(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto(); + } + + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser + PARSER = new com.google.protobuf.AbstractParser() { + @java.lang.Override + public SubmitWorkResponseProto parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new SubmitWorkResponseProto(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SubmitWorkResponseProto getDefaultInstanceForType() { + return DEFAULT_INSTANCE; } - // @@protoc_insertion_point(class_scope:SubmitWorkResponseProto) } - public interface SourceStateUpdatedRequestProtoOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface SourceStateUpdatedRequestProtoOrBuilder extends + // @@protoc_insertion_point(interface_extends:SourceStateUpdatedRequestProto) + com.google.protobuf.MessageOrBuilder { - // optional .QueryIdentifierProto query_identifier = 1; /** * optional .QueryIdentifierProto query_identifier = 1; + * @return Whether the queryIdentifier field is set. */ boolean hasQueryIdentifier(); /** * optional .QueryIdentifierProto query_identifier = 1; + * @return The queryIdentifier. */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto getQueryIdentifier(); /** @@ -13494,28 +15226,31 @@ public interface SourceStateUpdatedRequestProtoOrBuilder */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder getQueryIdentifierOrBuilder(); - // optional string src_name = 2; /** * optional string src_name = 2; + * @return Whether the srcName field is set. */ boolean hasSrcName(); /** * optional string src_name = 2; + * @return The srcName. */ java.lang.String getSrcName(); /** * optional string src_name = 2; + * @return The bytes for srcName. */ com.google.protobuf.ByteString getSrcNameBytes(); - // optional .SourceStateProto state = 3; /** * optional .SourceStateProto state = 3; + * @return Whether the state field is set. */ boolean hasState(); /** * optional .SourceStateProto state = 3; + * @return The state. */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateProto getState(); } @@ -13523,35 +15258,39 @@ public interface SourceStateUpdatedRequestProtoOrBuilder * Protobuf type {@code SourceStateUpdatedRequestProto} */ public static final class SourceStateUpdatedRequestProto extends - com.google.protobuf.GeneratedMessage - implements SourceStateUpdatedRequestProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:SourceStateUpdatedRequestProto) + SourceStateUpdatedRequestProtoOrBuilder { + private static final long serialVersionUID = 0L; // Use SourceStateUpdatedRequestProto.newBuilder() to construct. - private SourceStateUpdatedRequestProto(com.google.protobuf.GeneratedMessage.Builder builder) { + private SourceStateUpdatedRequestProto(com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private SourceStateUpdatedRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final SourceStateUpdatedRequestProto defaultInstance; - public static SourceStateUpdatedRequestProto getDefaultInstance() { - return defaultInstance; + private SourceStateUpdatedRequestProto() { + srcName_ = ""; + state_ = 1; } - public SourceStateUpdatedRequestProto getDefaultInstanceForType() { - return defaultInstance; + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance( + UnusedPrivateParameter unused) { + return new SourceStateUpdatedRequestProto(); } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private SourceStateUpdatedRequestProto( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); @@ -13563,16 +15302,9 @@ private SourceStateUpdatedRequestProto( case 0: done = true; break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } case 10: { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder subBuilder = null; - if (((bitField0_ & 0x00000001) == 0x00000001)) { + if (((bitField0_ & 0x00000001) != 0)) { subBuilder = queryIdentifier_.toBuilder(); } queryIdentifier_ = input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.PARSER, extensionRegistry); @@ -13584,18 +15316,27 @@ private SourceStateUpdatedRequestProto( break; } case 18: { + com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000002; - srcName_ = input.readBytes(); + srcName_ = bs; break; } case 24: { int rawValue = input.readEnum(); + @SuppressWarnings("deprecation") org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateProto value = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateProto.valueOf(rawValue); if (value == null) { unknownFields.mergeVarintField(3, rawValue); } else { bitField0_ |= 0x00000004; - state_ = value; + state_ = rawValue; + } + break; + } + default: { + if (!parseUnknownField( + input, unknownFields, extensionRegistry, tag)) { + done = true; } break; } @@ -13605,7 +15346,7 @@ private SourceStateUpdatedRequestProto( throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); @@ -13616,63 +15357,56 @@ private SourceStateUpdatedRequestProto( return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SourceStateUpdatedRequestProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SourceStateUpdatedRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public SourceStateUpdatedRequestProto parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new SourceStateUpdatedRequestProto(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - private int bitField0_; - // optional .QueryIdentifierProto query_identifier = 1; public static final int QUERY_IDENTIFIER_FIELD_NUMBER = 1; private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto queryIdentifier_; /** * optional .QueryIdentifierProto query_identifier = 1; + * @return Whether the queryIdentifier field is set. */ + @java.lang.Override public boolean hasQueryIdentifier() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional .QueryIdentifierProto query_identifier = 1; + * @return The queryIdentifier. */ + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto getQueryIdentifier() { - return queryIdentifier_; + return queryIdentifier_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance() : queryIdentifier_; } /** * optional .QueryIdentifierProto query_identifier = 1; */ + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder getQueryIdentifierOrBuilder() { - return queryIdentifier_; + return queryIdentifier_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance() : queryIdentifier_; } - // optional string src_name = 2; public static final int SRC_NAME_FIELD_NUMBER = 2; - private java.lang.Object srcName_; + private volatile java.lang.Object srcName_; /** * optional string src_name = 2; + * @return Whether the srcName field is set. */ + @java.lang.Override public boolean hasSrcName() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * optional string src_name = 2; + * @return The srcName. */ + @java.lang.Override public java.lang.String getSrcName() { java.lang.Object ref = srcName_; if (ref instanceof java.lang.String) { @@ -13689,7 +15423,9 @@ public java.lang.String getSrcName() { } /** * optional string src_name = 2; + * @return The bytes for srcName. */ + @java.lang.Override public com.google.protobuf.ByteString getSrcNameBytes() { java.lang.Object ref = srcName_; @@ -13704,81 +15440,73 @@ public java.lang.String getSrcName() { } } - // optional .SourceStateProto state = 3; public static final int STATE_FIELD_NUMBER = 3; - private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateProto state_; + private int state_; /** * optional .SourceStateProto state = 3; + * @return Whether the state field is set. */ - public boolean hasState() { - return ((bitField0_ & 0x00000004) == 0x00000004); + @java.lang.Override public boolean hasState() { + return ((bitField0_ & 0x00000004) != 0); } /** * optional .SourceStateProto state = 3; + * @return The state. */ - public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateProto getState() { - return state_; + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateProto getState() { + @SuppressWarnings("deprecation") + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateProto result = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateProto.valueOf(state_); + return result == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateProto.S_SUCCEEDED : result; } - private void initFields() { - queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance(); - srcName_ = ""; - state_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateProto.S_SUCCEEDED; - } private byte memoizedIsInitialized = -1; + @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } + @java.lang.Override public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); - if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeMessage(1, queryIdentifier_); + if (((bitField0_ & 0x00000001) != 0)) { + output.writeMessage(1, getQueryIdentifier()); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - output.writeBytes(2, getSrcNameBytes()); + if (((bitField0_ & 0x00000002) != 0)) { + com.google.protobuf.GeneratedMessageV3.writeString(output, 2, srcName_); } - if (((bitField0_ & 0x00000004) == 0x00000004)) { - output.writeEnum(3, state_.getNumber()); + if (((bitField0_ & 0x00000004) != 0)) { + output.writeEnum(3, state_); } - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; + @java.lang.Override public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { + if (((bitField0_ & 0x00000001) != 0)) { size += com.google.protobuf.CodedOutputStream - .computeMessageSize(1, queryIdentifier_); + .computeMessageSize(1, getQueryIdentifier()); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(2, getSrcNameBytes()); + if (((bitField0_ & 0x00000002) != 0)) { + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(2, srcName_); } - if (((bitField0_ & 0x00000004) == 0x00000004)) { + if (((bitField0_ & 0x00000004) != 0)) { size += com.google.protobuf.CodedOutputStream - .computeEnumSize(3, state_.getNumber()); + .computeEnumSize(3, state_); } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { @@ -13789,35 +15517,31 @@ public boolean equals(final java.lang.Object obj) { } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto) obj; - boolean result = true; - result = result && (hasQueryIdentifier() == other.hasQueryIdentifier()); + if (hasQueryIdentifier() != other.hasQueryIdentifier()) return false; if (hasQueryIdentifier()) { - result = result && getQueryIdentifier() - .equals(other.getQueryIdentifier()); + if (!getQueryIdentifier() + .equals(other.getQueryIdentifier())) return false; } - result = result && (hasSrcName() == other.hasSrcName()); + if (hasSrcName() != other.hasSrcName()) return false; if (hasSrcName()) { - result = result && getSrcName() - .equals(other.getSrcName()); + if (!getSrcName() + .equals(other.getSrcName())) return false; } - result = result && (hasState() == other.hasState()); + if (hasState() != other.hasState()) return false; if (hasState()) { - result = result && - (getState() == other.getState()); + if (state_ != other.state_) return false; } - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; + if (!unknownFields.equals(other.unknownFields)) return false; + return true; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); + hash = (19 * hash) + getDescriptor().hashCode(); if (hasQueryIdentifier()) { hash = (37 * hash) + QUERY_IDENTIFIER_FIELD_NUMBER; hash = (53 * hash) + getQueryIdentifier().hashCode(); @@ -13828,13 +15552,24 @@ public int hashCode() { } if (hasState()) { hash = (37 * hash) + STATE_FIELD_NUMBER; - hash = (53 * hash) + hashEnum(getState()); + hash = (53 * hash) + state_; } - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto parseFrom( + java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto parseFrom( + java.nio.ByteBuffer data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { @@ -13858,46 +15593,59 @@ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.So } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } + @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -13905,14 +15653,16 @@ protected Builder newBuilderForType( * Protobuf type {@code SourceStateUpdatedRequestProto} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder implements + // @@protoc_insertion_point(builder_implements:SourceStateUpdatedRequestProto) + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProtoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SourceStateUpdatedRequestProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SourceStateUpdatedRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -13925,47 +15675,44 @@ private Builder() { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { getQueryIdentifierFieldBuilder(); } } - private static Builder create() { - return new Builder(); - } - + @java.lang.Override public Builder clear() { super.clear(); if (queryIdentifierBuilder_ == null) { - queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance(); + queryIdentifier_ = null; } else { queryIdentifierBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); srcName_ = ""; bitField0_ = (bitField0_ & ~0x00000002); - state_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateProto.S_SUCCEEDED; + state_ = 1; bitField0_ = (bitField0_ & ~0x00000004); return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - + @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SourceStateUpdatedRequestProto_descriptor; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto.getDefaultInstance(); } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto result = buildPartial(); if (!result.isInitialized()) { @@ -13974,23 +15721,24 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceSta return result; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { + if (((from_bitField0_ & 0x00000001) != 0)) { + if (queryIdentifierBuilder_ == null) { + result.queryIdentifier_ = queryIdentifier_; + } else { + result.queryIdentifier_ = queryIdentifierBuilder_.build(); + } to_bitField0_ |= 0x00000001; } - if (queryIdentifierBuilder_ == null) { - result.queryIdentifier_ = queryIdentifier_; - } else { - result.queryIdentifier_ = queryIdentifierBuilder_.build(); - } - if (((from_bitField0_ & 0x00000002) == 0x00000002)) { + if (((from_bitField0_ & 0x00000002) != 0)) { to_bitField0_ |= 0x00000002; } result.srcName_ = srcName_; - if (((from_bitField0_ & 0x00000004) == 0x00000004)) { + if (((from_bitField0_ & 0x00000004) != 0)) { to_bitField0_ |= 0x00000004; } result.state_ = state_; @@ -13999,6 +15747,39 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceSta return result; } + @java.lang.Override + public Builder clone() { + return super.clone(); + } + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.setField(field, value); + } + @java.lang.Override + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + @java.lang.Override + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.addRepeatedField(field, value); + } + @java.lang.Override public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto)other); @@ -14021,14 +15802,17 @@ public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtoc if (other.hasState()) { setState(other.getState()); } - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } + @java.lang.Override public final boolean isInitialized() { return true; } + @java.lang.Override public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) @@ -14038,7 +15822,7 @@ public Builder mergeFrom( parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -14048,22 +15832,23 @@ public Builder mergeFrom( } private int bitField0_; - // optional .QueryIdentifierProto query_identifier = 1; - private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance(); - private com.google.protobuf.SingleFieldBuilder< + private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto queryIdentifier_; + private com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder> queryIdentifierBuilder_; /** * optional .QueryIdentifierProto query_identifier = 1; + * @return Whether the queryIdentifier field is set. */ public boolean hasQueryIdentifier() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional .QueryIdentifierProto query_identifier = 1; + * @return The queryIdentifier. */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto getQueryIdentifier() { if (queryIdentifierBuilder_ == null) { - return queryIdentifier_; + return queryIdentifier_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance() : queryIdentifier_; } else { return queryIdentifierBuilder_.getMessage(); } @@ -14103,7 +15888,8 @@ public Builder setQueryIdentifier( */ public Builder mergeQueryIdentifier(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto value) { if (queryIdentifierBuilder_ == null) { - if (((bitField0_ & 0x00000001) == 0x00000001) && + if (((bitField0_ & 0x00000001) != 0) && + queryIdentifier_ != null && queryIdentifier_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance()) { queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.newBuilder(queryIdentifier_).mergeFrom(value).buildPartial(); @@ -14122,7 +15908,7 @@ public Builder mergeQueryIdentifier(org.apache.hadoop.hive.llap.daemon.rpc.LlapD */ public Builder clearQueryIdentifier() { if (queryIdentifierBuilder_ == null) { - queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance(); + queryIdentifier_ = null; onChanged(); } else { queryIdentifierBuilder_.clear(); @@ -14145,19 +15931,20 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIden if (queryIdentifierBuilder_ != null) { return queryIdentifierBuilder_.getMessageOrBuilder(); } else { - return queryIdentifier_; + return queryIdentifier_ == null ? + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance() : queryIdentifier_; } } /** * optional .QueryIdentifierProto query_identifier = 1; */ - private com.google.protobuf.SingleFieldBuilder< + private com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder> getQueryIdentifierFieldBuilder() { if (queryIdentifierBuilder_ == null) { - queryIdentifierBuilder_ = new com.google.protobuf.SingleFieldBuilder< + queryIdentifierBuilder_ = new com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder>( - queryIdentifier_, + getQueryIdentifier(), getParentForChildren(), isClean()); queryIdentifier_ = null; @@ -14165,23 +15952,27 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIden return queryIdentifierBuilder_; } - // optional string src_name = 2; private java.lang.Object srcName_ = ""; /** * optional string src_name = 2; + * @return Whether the srcName field is set. */ public boolean hasSrcName() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * optional string src_name = 2; + * @return The srcName. */ public java.lang.String getSrcName() { java.lang.Object ref = srcName_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - srcName_ = s; + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + srcName_ = s; + } return s; } else { return (java.lang.String) ref; @@ -14189,6 +15980,7 @@ public java.lang.String getSrcName() { } /** * optional string src_name = 2; + * @return The bytes for srcName. */ public com.google.protobuf.ByteString getSrcNameBytes() { @@ -14205,6 +15997,8 @@ public java.lang.String getSrcName() { } /** * optional string src_name = 2; + * @param value The srcName to set. + * @return This builder for chaining. */ public Builder setSrcName( java.lang.String value) { @@ -14218,6 +16012,7 @@ public Builder setSrcName( } /** * optional string src_name = 2; + * @return This builder for chaining. */ public Builder clearSrcName() { bitField0_ = (bitField0_ & ~0x00000002); @@ -14227,6 +16022,8 @@ public Builder clearSrcName() { } /** * optional string src_name = 2; + * @param value The bytes for srcName to set. + * @return This builder for chaining. */ public Builder setSrcNameBytes( com.google.protobuf.ByteString value) { @@ -14239,89 +16036,140 @@ public Builder setSrcNameBytes( return this; } - // optional .SourceStateProto state = 3; - private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateProto state_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateProto.S_SUCCEEDED; + private int state_ = 1; /** * optional .SourceStateProto state = 3; + * @return Whether the state field is set. */ - public boolean hasState() { - return ((bitField0_ & 0x00000004) == 0x00000004); + @java.lang.Override public boolean hasState() { + return ((bitField0_ & 0x00000004) != 0); } /** * optional .SourceStateProto state = 3; + * @return The state. */ + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateProto getState() { - return state_; + @SuppressWarnings("deprecation") + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateProto result = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateProto.valueOf(state_); + return result == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateProto.S_SUCCEEDED : result; } /** * optional .SourceStateProto state = 3; + * @param value The state to set. + * @return This builder for chaining. */ public Builder setState(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateProto value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000004; - state_ = value; + state_ = value.getNumber(); onChanged(); return this; } /** * optional .SourceStateProto state = 3; + * @return This builder for chaining. */ public Builder clearState() { bitField0_ = (bitField0_ & ~0x00000004); - state_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateProto.S_SUCCEEDED; + state_ = 1; onChanged(); return this; } + @java.lang.Override + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:SourceStateUpdatedRequestProto) } + // @@protoc_insertion_point(class_scope:SourceStateUpdatedRequestProto) + private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto DEFAULT_INSTANCE; static { - defaultInstance = new SourceStateUpdatedRequestProto(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto(); + } + + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser + PARSER = new com.google.protobuf.AbstractParser() { + @java.lang.Override + public SourceStateUpdatedRequestProto parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new SourceStateUpdatedRequestProto(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedRequestProto getDefaultInstanceForType() { + return DEFAULT_INSTANCE; } - // @@protoc_insertion_point(class_scope:SourceStateUpdatedRequestProto) } - public interface SourceStateUpdatedResponseProtoOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface SourceStateUpdatedResponseProtoOrBuilder extends + // @@protoc_insertion_point(interface_extends:SourceStateUpdatedResponseProto) + com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code SourceStateUpdatedResponseProto} */ public static final class SourceStateUpdatedResponseProto extends - com.google.protobuf.GeneratedMessage - implements SourceStateUpdatedResponseProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:SourceStateUpdatedResponseProto) + SourceStateUpdatedResponseProtoOrBuilder { + private static final long serialVersionUID = 0L; // Use SourceStateUpdatedResponseProto.newBuilder() to construct. - private SourceStateUpdatedResponseProto(com.google.protobuf.GeneratedMessage.Builder builder) { + private SourceStateUpdatedResponseProto(com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private SourceStateUpdatedResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final SourceStateUpdatedResponseProto defaultInstance; - public static SourceStateUpdatedResponseProto getDefaultInstance() { - return defaultInstance; + private SourceStateUpdatedResponseProto() { } - public SourceStateUpdatedResponseProto getDefaultInstanceForType() { - return defaultInstance; + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance( + UnusedPrivateParameter unused) { + return new SourceStateUpdatedResponseProto(); } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private SourceStateUpdatedResponseProto( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); try { @@ -14333,8 +16181,8 @@ private SourceStateUpdatedResponseProto( done = true; break; default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { + if (!parseUnknownField( + input, unknownFields, extensionRegistry, tag)) { done = true; } break; @@ -14345,7 +16193,7 @@ private SourceStateUpdatedResponseProto( throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); @@ -14356,63 +16204,42 @@ private SourceStateUpdatedResponseProto( return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SourceStateUpdatedResponseProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SourceStateUpdatedResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public SourceStateUpdatedResponseProto parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new SourceStateUpdatedResponseProto(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - - private void initFields() { - } private byte memoizedIsInitialized = -1; + @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } + @java.lang.Override public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; + @java.lang.Override public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { @@ -14423,25 +16250,33 @@ public boolean equals(final java.lang.Object obj) { } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto) obj; - boolean result = true; - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; + if (!unknownFields.equals(other.unknownFields)) return false; + return true; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (19 * hash) + getDescriptor().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto parseFrom( + java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto parseFrom( + java.nio.ByteBuffer data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { @@ -14465,46 +16300,59 @@ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.So } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } + @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -14512,14 +16360,16 @@ protected Builder newBuilderForType( * Protobuf type {@code SourceStateUpdatedResponseProto} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder implements + // @@protoc_insertion_point(builder_implements:SourceStateUpdatedResponseProto) + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProtoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SourceStateUpdatedResponseProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SourceStateUpdatedResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -14532,36 +16382,33 @@ private Builder() { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { } } - private static Builder create() { - return new Builder(); - } - + @java.lang.Override public Builder clear() { super.clear(); return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - + @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SourceStateUpdatedResponseProto_descriptor; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto.getDefaultInstance(); } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto result = buildPartial(); if (!result.isInitialized()) { @@ -14570,12 +16417,46 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceSta return result; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto(this); onBuilt(); return result; } + @java.lang.Override + public Builder clone() { + return super.clone(); + } + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.setField(field, value); + } + @java.lang.Override + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + @java.lang.Override + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.addRepeatedField(field, value); + } + @java.lang.Override public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto)other); @@ -14587,14 +16468,17 @@ public Builder mergeFrom(com.google.protobuf.Message other) { public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto other) { if (other == org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto.getDefaultInstance()) return this; - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } + @java.lang.Override public final boolean isInitialized() { return true; } + @java.lang.Override public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) @@ -14604,7 +16488,7 @@ public Builder mergeFrom( parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -14612,28 +16496,71 @@ public Builder mergeFrom( } return this; } + @java.lang.Override + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:SourceStateUpdatedResponseProto) } + // @@protoc_insertion_point(class_scope:SourceStateUpdatedResponseProto) + private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto DEFAULT_INSTANCE; static { - defaultInstance = new SourceStateUpdatedResponseProto(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto(); + } + + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser + PARSER = new com.google.protobuf.AbstractParser() { + @java.lang.Override + public SourceStateUpdatedResponseProto parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new SourceStateUpdatedResponseProto(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SourceStateUpdatedResponseProto getDefaultInstanceForType() { + return DEFAULT_INSTANCE; } - // @@protoc_insertion_point(class_scope:SourceStateUpdatedResponseProto) } - public interface QueryCompleteRequestProtoOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface QueryCompleteRequestProtoOrBuilder extends + // @@protoc_insertion_point(interface_extends:QueryCompleteRequestProto) + com.google.protobuf.MessageOrBuilder { - // optional .QueryIdentifierProto query_identifier = 1; /** * optional .QueryIdentifierProto query_identifier = 1; + * @return Whether the queryIdentifier field is set. */ boolean hasQueryIdentifier(); /** * optional .QueryIdentifierProto query_identifier = 1; + * @return The queryIdentifier. */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto getQueryIdentifier(); /** @@ -14641,13 +16568,14 @@ public interface QueryCompleteRequestProtoOrBuilder */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder getQueryIdentifierOrBuilder(); - // optional int64 delete_delay = 2 [default = 0]; /** * optional int64 delete_delay = 2 [default = 0]; + * @return Whether the deleteDelay field is set. */ boolean hasDeleteDelay(); /** * optional int64 delete_delay = 2 [default = 0]; + * @return The deleteDelay. */ long getDeleteDelay(); } @@ -14655,35 +16583,37 @@ public interface QueryCompleteRequestProtoOrBuilder * Protobuf type {@code QueryCompleteRequestProto} */ public static final class QueryCompleteRequestProto extends - com.google.protobuf.GeneratedMessage - implements QueryCompleteRequestProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:QueryCompleteRequestProto) + QueryCompleteRequestProtoOrBuilder { + private static final long serialVersionUID = 0L; // Use QueryCompleteRequestProto.newBuilder() to construct. - private QueryCompleteRequestProto(com.google.protobuf.GeneratedMessage.Builder builder) { + private QueryCompleteRequestProto(com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private QueryCompleteRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final QueryCompleteRequestProto defaultInstance; - public static QueryCompleteRequestProto getDefaultInstance() { - return defaultInstance; + private QueryCompleteRequestProto() { } - public QueryCompleteRequestProto getDefaultInstanceForType() { - return defaultInstance; + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance( + UnusedPrivateParameter unused) { + return new QueryCompleteRequestProto(); } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private QueryCompleteRequestProto( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); @@ -14695,16 +16625,9 @@ private QueryCompleteRequestProto( case 0: done = true; break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } case 10: { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder subBuilder = null; - if (((bitField0_ & 0x00000001) == 0x00000001)) { + if (((bitField0_ & 0x00000001) != 0)) { subBuilder = queryIdentifier_.toBuilder(); } queryIdentifier_ = input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.PARSER, extensionRegistry); @@ -14720,13 +16643,20 @@ private QueryCompleteRequestProto( deleteDelay_ = input.readInt64(); break; } + default: { + if (!parseUnknownField( + input, unknownFields, extensionRegistry, tag)) { + done = true; + } + break; + } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); @@ -14737,118 +16667,102 @@ private QueryCompleteRequestProto( return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryCompleteRequestProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryCompleteRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public QueryCompleteRequestProto parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new QueryCompleteRequestProto(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - private int bitField0_; - // optional .QueryIdentifierProto query_identifier = 1; public static final int QUERY_IDENTIFIER_FIELD_NUMBER = 1; private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto queryIdentifier_; /** * optional .QueryIdentifierProto query_identifier = 1; + * @return Whether the queryIdentifier field is set. */ + @java.lang.Override public boolean hasQueryIdentifier() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional .QueryIdentifierProto query_identifier = 1; + * @return The queryIdentifier. */ + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto getQueryIdentifier() { - return queryIdentifier_; + return queryIdentifier_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance() : queryIdentifier_; } /** * optional .QueryIdentifierProto query_identifier = 1; */ + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder getQueryIdentifierOrBuilder() { - return queryIdentifier_; + return queryIdentifier_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance() : queryIdentifier_; } - // optional int64 delete_delay = 2 [default = 0]; public static final int DELETE_DELAY_FIELD_NUMBER = 2; private long deleteDelay_; /** * optional int64 delete_delay = 2 [default = 0]; + * @return Whether the deleteDelay field is set. */ + @java.lang.Override public boolean hasDeleteDelay() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * optional int64 delete_delay = 2 [default = 0]; + * @return The deleteDelay. */ + @java.lang.Override public long getDeleteDelay() { return deleteDelay_; } - private void initFields() { - queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance(); - deleteDelay_ = 0L; - } private byte memoizedIsInitialized = -1; + @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } + @java.lang.Override public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); - if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeMessage(1, queryIdentifier_); + if (((bitField0_ & 0x00000001) != 0)) { + output.writeMessage(1, getQueryIdentifier()); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { + if (((bitField0_ & 0x00000002) != 0)) { output.writeInt64(2, deleteDelay_); } - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; + @java.lang.Override public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { + if (((bitField0_ & 0x00000001) != 0)) { size += com.google.protobuf.CodedOutputStream - .computeMessageSize(1, queryIdentifier_); + .computeMessageSize(1, getQueryIdentifier()); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { + if (((bitField0_ & 0x00000002) != 0)) { size += com.google.protobuf.CodedOutputStream .computeInt64Size(2, deleteDelay_); } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { @@ -14859,43 +16773,52 @@ public boolean equals(final java.lang.Object obj) { } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto) obj; - boolean result = true; - result = result && (hasQueryIdentifier() == other.hasQueryIdentifier()); + if (hasQueryIdentifier() != other.hasQueryIdentifier()) return false; if (hasQueryIdentifier()) { - result = result && getQueryIdentifier() - .equals(other.getQueryIdentifier()); + if (!getQueryIdentifier() + .equals(other.getQueryIdentifier())) return false; } - result = result && (hasDeleteDelay() == other.hasDeleteDelay()); + if (hasDeleteDelay() != other.hasDeleteDelay()) return false; if (hasDeleteDelay()) { - result = result && (getDeleteDelay() - == other.getDeleteDelay()); + if (getDeleteDelay() + != other.getDeleteDelay()) return false; } - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; + if (!unknownFields.equals(other.unknownFields)) return false; + return true; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); + hash = (19 * hash) + getDescriptor().hashCode(); if (hasQueryIdentifier()) { hash = (37 * hash) + QUERY_IDENTIFIER_FIELD_NUMBER; hash = (53 * hash) + getQueryIdentifier().hashCode(); } if (hasDeleteDelay()) { hash = (37 * hash) + DELETE_DELAY_FIELD_NUMBER; - hash = (53 * hash) + hashLong(getDeleteDelay()); + hash = (53 * hash) + com.google.protobuf.Internal.hashLong( + getDeleteDelay()); } - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto parseFrom( + java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto parseFrom( + java.nio.ByteBuffer data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { @@ -14919,46 +16842,59 @@ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.Qu } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } + @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -14966,14 +16902,16 @@ protected Builder newBuilderForType( * Protobuf type {@code QueryCompleteRequestProto} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder implements + // @@protoc_insertion_point(builder_implements:QueryCompleteRequestProto) + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProtoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryCompleteRequestProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryCompleteRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -14986,23 +16924,21 @@ private Builder() { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { getQueryIdentifierFieldBuilder(); } } - private static Builder create() { - return new Builder(); - } - + @java.lang.Override public Builder clear() { super.clear(); if (queryIdentifierBuilder_ == null) { - queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance(); + queryIdentifier_ = null; } else { queryIdentifierBuilder_.clear(); } @@ -15012,19 +16948,18 @@ public Builder clear() { return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - + @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryCompleteRequestProto_descriptor; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto.getDefaultInstance(); } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto result = buildPartial(); if (!result.isInitialized()) { @@ -15033,27 +16968,61 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryComp return result; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { + if (((from_bitField0_ & 0x00000001) != 0)) { + if (queryIdentifierBuilder_ == null) { + result.queryIdentifier_ = queryIdentifier_; + } else { + result.queryIdentifier_ = queryIdentifierBuilder_.build(); + } to_bitField0_ |= 0x00000001; } - if (queryIdentifierBuilder_ == null) { - result.queryIdentifier_ = queryIdentifier_; - } else { - result.queryIdentifier_ = queryIdentifierBuilder_.build(); - } - if (((from_bitField0_ & 0x00000002) == 0x00000002)) { + if (((from_bitField0_ & 0x00000002) != 0)) { + result.deleteDelay_ = deleteDelay_; to_bitField0_ |= 0x00000002; } - result.deleteDelay_ = deleteDelay_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } + @java.lang.Override + public Builder clone() { + return super.clone(); + } + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.setField(field, value); + } + @java.lang.Override + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + @java.lang.Override + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.addRepeatedField(field, value); + } + @java.lang.Override public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto)other); @@ -15071,14 +17040,17 @@ public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtoc if (other.hasDeleteDelay()) { setDeleteDelay(other.getDeleteDelay()); } - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } + @java.lang.Override public final boolean isInitialized() { return true; } + @java.lang.Override public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) @@ -15088,7 +17060,7 @@ public Builder mergeFrom( parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -15098,22 +17070,23 @@ public Builder mergeFrom( } private int bitField0_; - // optional .QueryIdentifierProto query_identifier = 1; - private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance(); - private com.google.protobuf.SingleFieldBuilder< + private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto queryIdentifier_; + private com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder> queryIdentifierBuilder_; /** * optional .QueryIdentifierProto query_identifier = 1; + * @return Whether the queryIdentifier field is set. */ public boolean hasQueryIdentifier() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional .QueryIdentifierProto query_identifier = 1; + * @return The queryIdentifier. */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto getQueryIdentifier() { if (queryIdentifierBuilder_ == null) { - return queryIdentifier_; + return queryIdentifier_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance() : queryIdentifier_; } else { return queryIdentifierBuilder_.getMessage(); } @@ -15153,7 +17126,8 @@ public Builder setQueryIdentifier( */ public Builder mergeQueryIdentifier(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto value) { if (queryIdentifierBuilder_ == null) { - if (((bitField0_ & 0x00000001) == 0x00000001) && + if (((bitField0_ & 0x00000001) != 0) && + queryIdentifier_ != null && queryIdentifier_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance()) { queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.newBuilder(queryIdentifier_).mergeFrom(value).buildPartial(); @@ -15172,7 +17146,7 @@ public Builder mergeQueryIdentifier(org.apache.hadoop.hive.llap.daemon.rpc.LlapD */ public Builder clearQueryIdentifier() { if (queryIdentifierBuilder_ == null) { - queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance(); + queryIdentifier_ = null; onChanged(); } else { queryIdentifierBuilder_.clear(); @@ -15195,19 +17169,20 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIden if (queryIdentifierBuilder_ != null) { return queryIdentifierBuilder_.getMessageOrBuilder(); } else { - return queryIdentifier_; + return queryIdentifier_ == null ? + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance() : queryIdentifier_; } } /** * optional .QueryIdentifierProto query_identifier = 1; */ - private com.google.protobuf.SingleFieldBuilder< + private com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder> getQueryIdentifierFieldBuilder() { if (queryIdentifierBuilder_ == null) { - queryIdentifierBuilder_ = new com.google.protobuf.SingleFieldBuilder< + queryIdentifierBuilder_ = new com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder>( - queryIdentifier_, + getQueryIdentifier(), getParentForChildren(), isClean()); queryIdentifier_ = null; @@ -15215,22 +17190,27 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIden return queryIdentifierBuilder_; } - // optional int64 delete_delay = 2 [default = 0]; private long deleteDelay_ ; /** * optional int64 delete_delay = 2 [default = 0]; + * @return Whether the deleteDelay field is set. */ + @java.lang.Override public boolean hasDeleteDelay() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * optional int64 delete_delay = 2 [default = 0]; + * @return The deleteDelay. */ + @java.lang.Override public long getDeleteDelay() { return deleteDelay_; } /** * optional int64 delete_delay = 2 [default = 0]; + * @param value The deleteDelay to set. + * @return This builder for chaining. */ public Builder setDeleteDelay(long value) { bitField0_ |= 0x00000002; @@ -15240,6 +17220,7 @@ public Builder setDeleteDelay(long value) { } /** * optional int64 delete_delay = 2 [default = 0]; + * @return This builder for chaining. */ public Builder clearDeleteDelay() { bitField0_ = (bitField0_ & ~0x00000002); @@ -15247,54 +17228,98 @@ public Builder clearDeleteDelay() { onChanged(); return this; } + @java.lang.Override + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:QueryCompleteRequestProto) } + // @@protoc_insertion_point(class_scope:QueryCompleteRequestProto) + private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto DEFAULT_INSTANCE; static { - defaultInstance = new QueryCompleteRequestProto(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto(); + } + + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser + PARSER = new com.google.protobuf.AbstractParser() { + @java.lang.Override + public QueryCompleteRequestProto parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new QueryCompleteRequestProto(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteRequestProto getDefaultInstanceForType() { + return DEFAULT_INSTANCE; } - // @@protoc_insertion_point(class_scope:QueryCompleteRequestProto) } - public interface QueryCompleteResponseProtoOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface QueryCompleteResponseProtoOrBuilder extends + // @@protoc_insertion_point(interface_extends:QueryCompleteResponseProto) + com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code QueryCompleteResponseProto} */ public static final class QueryCompleteResponseProto extends - com.google.protobuf.GeneratedMessage - implements QueryCompleteResponseProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:QueryCompleteResponseProto) + QueryCompleteResponseProtoOrBuilder { + private static final long serialVersionUID = 0L; // Use QueryCompleteResponseProto.newBuilder() to construct. - private QueryCompleteResponseProto(com.google.protobuf.GeneratedMessage.Builder builder) { + private QueryCompleteResponseProto(com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private QueryCompleteResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final QueryCompleteResponseProto defaultInstance; - public static QueryCompleteResponseProto getDefaultInstance() { - return defaultInstance; + private QueryCompleteResponseProto() { } - public QueryCompleteResponseProto getDefaultInstanceForType() { - return defaultInstance; + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance( + UnusedPrivateParameter unused) { + return new QueryCompleteResponseProto(); } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private QueryCompleteResponseProto( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); try { @@ -15306,8 +17331,8 @@ private QueryCompleteResponseProto( done = true; break; default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { + if (!parseUnknownField( + input, unknownFields, extensionRegistry, tag)) { done = true; } break; @@ -15318,7 +17343,7 @@ private QueryCompleteResponseProto( throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); @@ -15329,63 +17354,42 @@ private QueryCompleteResponseProto( return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryCompleteResponseProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryCompleteResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public QueryCompleteResponseProto parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new QueryCompleteResponseProto(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - - private void initFields() { - } private byte memoizedIsInitialized = -1; + @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } + @java.lang.Override public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; + @java.lang.Override public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { @@ -15396,25 +17400,33 @@ public boolean equals(final java.lang.Object obj) { } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto) obj; - boolean result = true; - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; + if (!unknownFields.equals(other.unknownFields)) return false; + return true; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (19 * hash) + getDescriptor().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto parseFrom( + java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto parseFrom( + java.nio.ByteBuffer data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { @@ -15438,46 +17450,59 @@ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.Qu } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } + @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -15485,14 +17510,16 @@ protected Builder newBuilderForType( * Protobuf type {@code QueryCompleteResponseProto} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder implements + // @@protoc_insertion_point(builder_implements:QueryCompleteResponseProto) + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProtoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryCompleteResponseProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryCompleteResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -15505,36 +17532,33 @@ private Builder() { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { } } - private static Builder create() { - return new Builder(); - } - + @java.lang.Override public Builder clear() { super.clear(); return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - + @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_QueryCompleteResponseProto_descriptor; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto.getDefaultInstance(); } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto result = buildPartial(); if (!result.isInitialized()) { @@ -15543,12 +17567,46 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryComp return result; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto(this); onBuilt(); return result; } + @java.lang.Override + public Builder clone() { + return super.clone(); + } + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.setField(field, value); + } + @java.lang.Override + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + @java.lang.Override + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.addRepeatedField(field, value); + } + @java.lang.Override public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto)other); @@ -15560,14 +17618,17 @@ public Builder mergeFrom(com.google.protobuf.Message other) { public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto other) { if (other == org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto.getDefaultInstance()) return this; - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } + @java.lang.Override public final boolean isInitialized() { return true; } + @java.lang.Override public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) @@ -15577,7 +17638,7 @@ public Builder mergeFrom( parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -15585,28 +17646,71 @@ public Builder mergeFrom( } return this; } + @java.lang.Override + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:QueryCompleteResponseProto) } + // @@protoc_insertion_point(class_scope:QueryCompleteResponseProto) + private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto DEFAULT_INSTANCE; static { - defaultInstance = new QueryCompleteResponseProto(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto(); } - // @@protoc_insertion_point(class_scope:QueryCompleteResponseProto) - } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser + PARSER = new com.google.protobuf.AbstractParser() { + @java.lang.Override + public QueryCompleteResponseProto parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new QueryCompleteResponseProto(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryCompleteResponseProto getDefaultInstanceForType() { + return DEFAULT_INSTANCE; + } - public interface TerminateFragmentRequestProtoOrBuilder - extends com.google.protobuf.MessageOrBuilder { + } + + public interface TerminateFragmentRequestProtoOrBuilder extends + // @@protoc_insertion_point(interface_extends:TerminateFragmentRequestProto) + com.google.protobuf.MessageOrBuilder { - // optional .QueryIdentifierProto query_identifier = 1; /** * optional .QueryIdentifierProto query_identifier = 1; + * @return Whether the queryIdentifier field is set. */ boolean hasQueryIdentifier(); /** * optional .QueryIdentifierProto query_identifier = 1; + * @return The queryIdentifier. */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto getQueryIdentifier(); /** @@ -15614,17 +17718,19 @@ public interface TerminateFragmentRequestProtoOrBuilder */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder getQueryIdentifierOrBuilder(); - // optional string fragment_identifier_string = 2; /** * optional string fragment_identifier_string = 2; + * @return Whether the fragmentIdentifierString field is set. */ boolean hasFragmentIdentifierString(); /** * optional string fragment_identifier_string = 2; + * @return The fragmentIdentifierString. */ java.lang.String getFragmentIdentifierString(); /** * optional string fragment_identifier_string = 2; + * @return The bytes for fragmentIdentifierString. */ com.google.protobuf.ByteString getFragmentIdentifierStringBytes(); @@ -15633,35 +17739,38 @@ public interface TerminateFragmentRequestProtoOrBuilder * Protobuf type {@code TerminateFragmentRequestProto} */ public static final class TerminateFragmentRequestProto extends - com.google.protobuf.GeneratedMessage - implements TerminateFragmentRequestProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:TerminateFragmentRequestProto) + TerminateFragmentRequestProtoOrBuilder { + private static final long serialVersionUID = 0L; // Use TerminateFragmentRequestProto.newBuilder() to construct. - private TerminateFragmentRequestProto(com.google.protobuf.GeneratedMessage.Builder builder) { + private TerminateFragmentRequestProto(com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private TerminateFragmentRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final TerminateFragmentRequestProto defaultInstance; - public static TerminateFragmentRequestProto getDefaultInstance() { - return defaultInstance; + private TerminateFragmentRequestProto() { + fragmentIdentifierString_ = ""; } - public TerminateFragmentRequestProto getDefaultInstanceForType() { - return defaultInstance; + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance( + UnusedPrivateParameter unused) { + return new TerminateFragmentRequestProto(); } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private TerminateFragmentRequestProto( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); @@ -15673,16 +17782,9 @@ private TerminateFragmentRequestProto( case 0: done = true; break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } case 10: { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder subBuilder = null; - if (((bitField0_ & 0x00000001) == 0x00000001)) { + if (((bitField0_ & 0x00000001) != 0)) { subBuilder = queryIdentifier_.toBuilder(); } queryIdentifier_ = input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.PARSER, extensionRegistry); @@ -15694,8 +17796,16 @@ private TerminateFragmentRequestProto( break; } case 18: { + com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000002; - fragmentIdentifierString_ = input.readBytes(); + fragmentIdentifierString_ = bs; + break; + } + default: { + if (!parseUnknownField( + input, unknownFields, extensionRegistry, tag)) { + done = true; + } break; } } @@ -15704,7 +17814,7 @@ private TerminateFragmentRequestProto( throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); @@ -15715,63 +17825,56 @@ private TerminateFragmentRequestProto( return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_TerminateFragmentRequestProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_TerminateFragmentRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public TerminateFragmentRequestProto parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new TerminateFragmentRequestProto(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - private int bitField0_; - // optional .QueryIdentifierProto query_identifier = 1; public static final int QUERY_IDENTIFIER_FIELD_NUMBER = 1; private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto queryIdentifier_; /** * optional .QueryIdentifierProto query_identifier = 1; + * @return Whether the queryIdentifier field is set. */ + @java.lang.Override public boolean hasQueryIdentifier() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional .QueryIdentifierProto query_identifier = 1; + * @return The queryIdentifier. */ + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto getQueryIdentifier() { - return queryIdentifier_; + return queryIdentifier_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance() : queryIdentifier_; } /** * optional .QueryIdentifierProto query_identifier = 1; */ + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder getQueryIdentifierOrBuilder() { - return queryIdentifier_; + return queryIdentifier_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance() : queryIdentifier_; } - // optional string fragment_identifier_string = 2; public static final int FRAGMENT_IDENTIFIER_STRING_FIELD_NUMBER = 2; - private java.lang.Object fragmentIdentifierString_; + private volatile java.lang.Object fragmentIdentifierString_; /** * optional string fragment_identifier_string = 2; + * @return Whether the fragmentIdentifierString field is set. */ + @java.lang.Override public boolean hasFragmentIdentifierString() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * optional string fragment_identifier_string = 2; + * @return The fragmentIdentifierString. */ + @java.lang.Override public java.lang.String getFragmentIdentifierString() { java.lang.Object ref = fragmentIdentifierString_; if (ref instanceof java.lang.String) { @@ -15788,7 +17891,9 @@ public java.lang.String getFragmentIdentifierString() { } /** * optional string fragment_identifier_string = 2; + * @return The bytes for fragmentIdentifierString. */ + @java.lang.Override public com.google.protobuf.ByteString getFragmentIdentifierStringBytes() { java.lang.Object ref = fragmentIdentifierString_; @@ -15803,57 +17908,47 @@ public java.lang.String getFragmentIdentifierString() { } } - private void initFields() { - queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance(); - fragmentIdentifierString_ = ""; - } private byte memoizedIsInitialized = -1; + @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } + @java.lang.Override public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); - if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeMessage(1, queryIdentifier_); + if (((bitField0_ & 0x00000001) != 0)) { + output.writeMessage(1, getQueryIdentifier()); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - output.writeBytes(2, getFragmentIdentifierStringBytes()); + if (((bitField0_ & 0x00000002) != 0)) { + com.google.protobuf.GeneratedMessageV3.writeString(output, 2, fragmentIdentifierString_); } - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; + @java.lang.Override public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { + if (((bitField0_ & 0x00000001) != 0)) { size += com.google.protobuf.CodedOutputStream - .computeMessageSize(1, queryIdentifier_); + .computeMessageSize(1, getQueryIdentifier()); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(2, getFragmentIdentifierStringBytes()); + if (((bitField0_ & 0x00000002) != 0)) { + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(2, fragmentIdentifierString_); } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { @@ -15864,30 +17959,27 @@ public boolean equals(final java.lang.Object obj) { } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto) obj; - boolean result = true; - result = result && (hasQueryIdentifier() == other.hasQueryIdentifier()); + if (hasQueryIdentifier() != other.hasQueryIdentifier()) return false; if (hasQueryIdentifier()) { - result = result && getQueryIdentifier() - .equals(other.getQueryIdentifier()); + if (!getQueryIdentifier() + .equals(other.getQueryIdentifier())) return false; } - result = result && (hasFragmentIdentifierString() == other.hasFragmentIdentifierString()); + if (hasFragmentIdentifierString() != other.hasFragmentIdentifierString()) return false; if (hasFragmentIdentifierString()) { - result = result && getFragmentIdentifierString() - .equals(other.getFragmentIdentifierString()); + if (!getFragmentIdentifierString() + .equals(other.getFragmentIdentifierString())) return false; } - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; + if (!unknownFields.equals(other.unknownFields)) return false; + return true; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); + hash = (19 * hash) + getDescriptor().hashCode(); if (hasQueryIdentifier()) { hash = (37 * hash) + QUERY_IDENTIFIER_FIELD_NUMBER; hash = (53 * hash) + getQueryIdentifier().hashCode(); @@ -15896,11 +17988,22 @@ public int hashCode() { hash = (37 * hash) + FRAGMENT_IDENTIFIER_STRING_FIELD_NUMBER; hash = (53 * hash) + getFragmentIdentifierString().hashCode(); } - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto parseFrom( + java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto parseFrom( + java.nio.ByteBuffer data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { @@ -15924,46 +18027,59 @@ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.Te } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } + @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -15971,14 +18087,16 @@ protected Builder newBuilderForType( * Protobuf type {@code TerminateFragmentRequestProto} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder implements + // @@protoc_insertion_point(builder_implements:TerminateFragmentRequestProto) + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProtoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_TerminateFragmentRequestProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_TerminateFragmentRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -15991,23 +18109,21 @@ private Builder() { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { getQueryIdentifierFieldBuilder(); } } - private static Builder create() { - return new Builder(); - } - + @java.lang.Override public Builder clear() { super.clear(); if (queryIdentifierBuilder_ == null) { - queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance(); + queryIdentifier_ = null; } else { queryIdentifierBuilder_.clear(); } @@ -16017,19 +18133,18 @@ public Builder clear() { return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - + @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_TerminateFragmentRequestProto_descriptor; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto.getDefaultInstance(); } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto result = buildPartial(); if (!result.isInitialized()) { @@ -16038,19 +18153,20 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.Terminate return result; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { + if (((from_bitField0_ & 0x00000001) != 0)) { + if (queryIdentifierBuilder_ == null) { + result.queryIdentifier_ = queryIdentifier_; + } else { + result.queryIdentifier_ = queryIdentifierBuilder_.build(); + } to_bitField0_ |= 0x00000001; } - if (queryIdentifierBuilder_ == null) { - result.queryIdentifier_ = queryIdentifier_; - } else { - result.queryIdentifier_ = queryIdentifierBuilder_.build(); - } - if (((from_bitField0_ & 0x00000002) == 0x00000002)) { + if (((from_bitField0_ & 0x00000002) != 0)) { to_bitField0_ |= 0x00000002; } result.fragmentIdentifierString_ = fragmentIdentifierString_; @@ -16059,6 +18175,39 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.Terminate return result; } + @java.lang.Override + public Builder clone() { + return super.clone(); + } + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.setField(field, value); + } + @java.lang.Override + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + @java.lang.Override + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.addRepeatedField(field, value); + } + @java.lang.Override public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto)other); @@ -16078,14 +18227,17 @@ public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtoc fragmentIdentifierString_ = other.fragmentIdentifierString_; onChanged(); } - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } + @java.lang.Override public final boolean isInitialized() { return true; } + @java.lang.Override public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) @@ -16095,7 +18247,7 @@ public Builder mergeFrom( parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -16105,22 +18257,23 @@ public Builder mergeFrom( } private int bitField0_; - // optional .QueryIdentifierProto query_identifier = 1; - private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance(); - private com.google.protobuf.SingleFieldBuilder< + private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto queryIdentifier_; + private com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder> queryIdentifierBuilder_; /** * optional .QueryIdentifierProto query_identifier = 1; + * @return Whether the queryIdentifier field is set. */ public boolean hasQueryIdentifier() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional .QueryIdentifierProto query_identifier = 1; + * @return The queryIdentifier. */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto getQueryIdentifier() { if (queryIdentifierBuilder_ == null) { - return queryIdentifier_; + return queryIdentifier_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance() : queryIdentifier_; } else { return queryIdentifierBuilder_.getMessage(); } @@ -16160,7 +18313,8 @@ public Builder setQueryIdentifier( */ public Builder mergeQueryIdentifier(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto value) { if (queryIdentifierBuilder_ == null) { - if (((bitField0_ & 0x00000001) == 0x00000001) && + if (((bitField0_ & 0x00000001) != 0) && + queryIdentifier_ != null && queryIdentifier_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance()) { queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.newBuilder(queryIdentifier_).mergeFrom(value).buildPartial(); @@ -16179,7 +18333,7 @@ public Builder mergeQueryIdentifier(org.apache.hadoop.hive.llap.daemon.rpc.LlapD */ public Builder clearQueryIdentifier() { if (queryIdentifierBuilder_ == null) { - queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance(); + queryIdentifier_ = null; onChanged(); } else { queryIdentifierBuilder_.clear(); @@ -16202,19 +18356,20 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIden if (queryIdentifierBuilder_ != null) { return queryIdentifierBuilder_.getMessageOrBuilder(); } else { - return queryIdentifier_; + return queryIdentifier_ == null ? + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance() : queryIdentifier_; } } /** * optional .QueryIdentifierProto query_identifier = 1; */ - private com.google.protobuf.SingleFieldBuilder< + private com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder> getQueryIdentifierFieldBuilder() { if (queryIdentifierBuilder_ == null) { - queryIdentifierBuilder_ = new com.google.protobuf.SingleFieldBuilder< + queryIdentifierBuilder_ = new com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder>( - queryIdentifier_, + getQueryIdentifier(), getParentForChildren(), isClean()); queryIdentifier_ = null; @@ -16222,23 +18377,27 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIden return queryIdentifierBuilder_; } - // optional string fragment_identifier_string = 2; private java.lang.Object fragmentIdentifierString_ = ""; /** * optional string fragment_identifier_string = 2; + * @return Whether the fragmentIdentifierString field is set. */ public boolean hasFragmentIdentifierString() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * optional string fragment_identifier_string = 2; + * @return The fragmentIdentifierString. */ public java.lang.String getFragmentIdentifierString() { java.lang.Object ref = fragmentIdentifierString_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - fragmentIdentifierString_ = s; + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + fragmentIdentifierString_ = s; + } return s; } else { return (java.lang.String) ref; @@ -16246,6 +18405,7 @@ public java.lang.String getFragmentIdentifierString() { } /** * optional string fragment_identifier_string = 2; + * @return The bytes for fragmentIdentifierString. */ public com.google.protobuf.ByteString getFragmentIdentifierStringBytes() { @@ -16262,6 +18422,8 @@ public java.lang.String getFragmentIdentifierString() { } /** * optional string fragment_identifier_string = 2; + * @param value The fragmentIdentifierString to set. + * @return This builder for chaining. */ public Builder setFragmentIdentifierString( java.lang.String value) { @@ -16275,6 +18437,7 @@ public Builder setFragmentIdentifierString( } /** * optional string fragment_identifier_string = 2; + * @return This builder for chaining. */ public Builder clearFragmentIdentifierString() { bitField0_ = (bitField0_ & ~0x00000002); @@ -16284,6 +18447,8 @@ public Builder clearFragmentIdentifierString() { } /** * optional string fragment_identifier_string = 2; + * @param value The bytes for fragmentIdentifierString to set. + * @return This builder for chaining. */ public Builder setFragmentIdentifierStringBytes( com.google.protobuf.ByteString value) { @@ -16295,54 +18460,98 @@ public Builder setFragmentIdentifierStringBytes( onChanged(); return this; } + @java.lang.Override + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:TerminateFragmentRequestProto) } + // @@protoc_insertion_point(class_scope:TerminateFragmentRequestProto) + private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto DEFAULT_INSTANCE; static { - defaultInstance = new TerminateFragmentRequestProto(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto(); + } + + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser + PARSER = new com.google.protobuf.AbstractParser() { + @java.lang.Override + public TerminateFragmentRequestProto parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new TerminateFragmentRequestProto(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentRequestProto getDefaultInstanceForType() { + return DEFAULT_INSTANCE; } - // @@protoc_insertion_point(class_scope:TerminateFragmentRequestProto) } - public interface TerminateFragmentResponseProtoOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface TerminateFragmentResponseProtoOrBuilder extends + // @@protoc_insertion_point(interface_extends:TerminateFragmentResponseProto) + com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code TerminateFragmentResponseProto} */ public static final class TerminateFragmentResponseProto extends - com.google.protobuf.GeneratedMessage - implements TerminateFragmentResponseProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:TerminateFragmentResponseProto) + TerminateFragmentResponseProtoOrBuilder { + private static final long serialVersionUID = 0L; // Use TerminateFragmentResponseProto.newBuilder() to construct. - private TerminateFragmentResponseProto(com.google.protobuf.GeneratedMessage.Builder builder) { + private TerminateFragmentResponseProto(com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private TerminateFragmentResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final TerminateFragmentResponseProto defaultInstance; - public static TerminateFragmentResponseProto getDefaultInstance() { - return defaultInstance; + private TerminateFragmentResponseProto() { } - public TerminateFragmentResponseProto getDefaultInstanceForType() { - return defaultInstance; + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance( + UnusedPrivateParameter unused) { + return new TerminateFragmentResponseProto(); } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private TerminateFragmentResponseProto( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); try { @@ -16354,8 +18563,8 @@ private TerminateFragmentResponseProto( done = true; break; default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { + if (!parseUnknownField( + input, unknownFields, extensionRegistry, tag)) { done = true; } break; @@ -16366,7 +18575,7 @@ private TerminateFragmentResponseProto( throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); @@ -16377,63 +18586,42 @@ private TerminateFragmentResponseProto( return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_TerminateFragmentResponseProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_TerminateFragmentResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public TerminateFragmentResponseProto parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new TerminateFragmentResponseProto(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - - private void initFields() { - } private byte memoizedIsInitialized = -1; + @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } + @java.lang.Override public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; + @java.lang.Override public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { @@ -16444,25 +18632,33 @@ public boolean equals(final java.lang.Object obj) { } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto) obj; - boolean result = true; - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; + if (!unknownFields.equals(other.unknownFields)) return false; + return true; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (19 * hash) + getDescriptor().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto parseFrom( + java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto parseFrom( + java.nio.ByteBuffer data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { @@ -16486,46 +18682,59 @@ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.Te } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } + @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -16533,14 +18742,16 @@ protected Builder newBuilderForType( * Protobuf type {@code TerminateFragmentResponseProto} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder implements + // @@protoc_insertion_point(builder_implements:TerminateFragmentResponseProto) + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProtoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_TerminateFragmentResponseProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_TerminateFragmentResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -16553,36 +18764,33 @@ private Builder() { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { } } - private static Builder create() { - return new Builder(); - } - + @java.lang.Override public Builder clear() { super.clear(); return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - + @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_TerminateFragmentResponseProto_descriptor; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto.getDefaultInstance(); } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto result = buildPartial(); if (!result.isInitialized()) { @@ -16591,12 +18799,46 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.Terminate return result; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto(this); onBuilt(); return result; } + @java.lang.Override + public Builder clone() { + return super.clone(); + } + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.setField(field, value); + } + @java.lang.Override + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + @java.lang.Override + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.addRepeatedField(field, value); + } + @java.lang.Override public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto)other); @@ -16608,14 +18850,17 @@ public Builder mergeFrom(com.google.protobuf.Message other) { public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto other) { if (other == org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto.getDefaultInstance()) return this; - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } + @java.lang.Override public final boolean isInitialized() { return true; } + @java.lang.Override public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) @@ -16625,7 +18870,7 @@ public Builder mergeFrom( parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -16633,28 +18878,71 @@ public Builder mergeFrom( } return this; } + @java.lang.Override + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:TerminateFragmentResponseProto) } + // @@protoc_insertion_point(class_scope:TerminateFragmentResponseProto) + private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto DEFAULT_INSTANCE; static { - defaultInstance = new TerminateFragmentResponseProto(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto(); + } + + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser + PARSER = new com.google.protobuf.AbstractParser() { + @java.lang.Override + public TerminateFragmentResponseProto parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new TerminateFragmentResponseProto(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TerminateFragmentResponseProto getDefaultInstanceForType() { + return DEFAULT_INSTANCE; } - // @@protoc_insertion_point(class_scope:TerminateFragmentResponseProto) } - public interface UpdateFragmentRequestProtoOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface UpdateFragmentRequestProtoOrBuilder extends + // @@protoc_insertion_point(interface_extends:UpdateFragmentRequestProto) + com.google.protobuf.MessageOrBuilder { - // optional .QueryIdentifierProto query_identifier = 1; /** * optional .QueryIdentifierProto query_identifier = 1; + * @return Whether the queryIdentifier field is set. */ boolean hasQueryIdentifier(); /** * optional .QueryIdentifierProto query_identifier = 1; + * @return The queryIdentifier. */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto getQueryIdentifier(); /** @@ -16662,28 +18950,31 @@ public interface UpdateFragmentRequestProtoOrBuilder */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder getQueryIdentifierOrBuilder(); - // optional string fragment_identifier_string = 2; /** * optional string fragment_identifier_string = 2; + * @return Whether the fragmentIdentifierString field is set. */ boolean hasFragmentIdentifierString(); /** * optional string fragment_identifier_string = 2; + * @return The fragmentIdentifierString. */ java.lang.String getFragmentIdentifierString(); /** * optional string fragment_identifier_string = 2; + * @return The bytes for fragmentIdentifierString. */ com.google.protobuf.ByteString getFragmentIdentifierStringBytes(); - // optional bool is_guaranteed = 3; /** * optional bool is_guaranteed = 3; + * @return Whether the isGuaranteed field is set. */ boolean hasIsGuaranteed(); /** * optional bool is_guaranteed = 3; + * @return The isGuaranteed. */ boolean getIsGuaranteed(); } @@ -16691,35 +18982,38 @@ public interface UpdateFragmentRequestProtoOrBuilder * Protobuf type {@code UpdateFragmentRequestProto} */ public static final class UpdateFragmentRequestProto extends - com.google.protobuf.GeneratedMessage - implements UpdateFragmentRequestProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:UpdateFragmentRequestProto) + UpdateFragmentRequestProtoOrBuilder { + private static final long serialVersionUID = 0L; // Use UpdateFragmentRequestProto.newBuilder() to construct. - private UpdateFragmentRequestProto(com.google.protobuf.GeneratedMessage.Builder builder) { + private UpdateFragmentRequestProto(com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private UpdateFragmentRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final UpdateFragmentRequestProto defaultInstance; - public static UpdateFragmentRequestProto getDefaultInstance() { - return defaultInstance; + private UpdateFragmentRequestProto() { + fragmentIdentifierString_ = ""; } - public UpdateFragmentRequestProto getDefaultInstanceForType() { - return defaultInstance; + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance( + UnusedPrivateParameter unused) { + return new UpdateFragmentRequestProto(); } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private UpdateFragmentRequestProto( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); @@ -16731,16 +19025,9 @@ private UpdateFragmentRequestProto( case 0: done = true; break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } case 10: { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder subBuilder = null; - if (((bitField0_ & 0x00000001) == 0x00000001)) { + if (((bitField0_ & 0x00000001) != 0)) { subBuilder = queryIdentifier_.toBuilder(); } queryIdentifier_ = input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.PARSER, extensionRegistry); @@ -16752,8 +19039,9 @@ private UpdateFragmentRequestProto( break; } case 18: { + com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000002; - fragmentIdentifierString_ = input.readBytes(); + fragmentIdentifierString_ = bs; break; } case 24: { @@ -16761,13 +19049,20 @@ private UpdateFragmentRequestProto( isGuaranteed_ = input.readBool(); break; } + default: { + if (!parseUnknownField( + input, unknownFields, extensionRegistry, tag)) { + done = true; + } + break; + } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); @@ -16778,63 +19073,56 @@ private UpdateFragmentRequestProto( return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_UpdateFragmentRequestProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_UpdateFragmentRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentRequestProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentRequestProto.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public UpdateFragmentRequestProto parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new UpdateFragmentRequestProto(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - private int bitField0_; - // optional .QueryIdentifierProto query_identifier = 1; public static final int QUERY_IDENTIFIER_FIELD_NUMBER = 1; private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto queryIdentifier_; /** * optional .QueryIdentifierProto query_identifier = 1; + * @return Whether the queryIdentifier field is set. */ + @java.lang.Override public boolean hasQueryIdentifier() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional .QueryIdentifierProto query_identifier = 1; + * @return The queryIdentifier. */ + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto getQueryIdentifier() { - return queryIdentifier_; + return queryIdentifier_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance() : queryIdentifier_; } /** * optional .QueryIdentifierProto query_identifier = 1; */ + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder getQueryIdentifierOrBuilder() { - return queryIdentifier_; + return queryIdentifier_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance() : queryIdentifier_; } - // optional string fragment_identifier_string = 2; public static final int FRAGMENT_IDENTIFIER_STRING_FIELD_NUMBER = 2; - private java.lang.Object fragmentIdentifierString_; + private volatile java.lang.Object fragmentIdentifierString_; /** * optional string fragment_identifier_string = 2; + * @return Whether the fragmentIdentifierString field is set. */ + @java.lang.Override public boolean hasFragmentIdentifierString() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * optional string fragment_identifier_string = 2; + * @return The fragmentIdentifierString. */ + @java.lang.Override public java.lang.String getFragmentIdentifierString() { java.lang.Object ref = fragmentIdentifierString_; if (ref instanceof java.lang.String) { @@ -16851,7 +19139,9 @@ public java.lang.String getFragmentIdentifierString() { } /** * optional string fragment_identifier_string = 2; + * @return The bytes for fragmentIdentifierString. */ + @java.lang.Override public com.google.protobuf.ByteString getFragmentIdentifierStringBytes() { java.lang.Object ref = fragmentIdentifierString_; @@ -16866,81 +19156,73 @@ public java.lang.String getFragmentIdentifierString() { } } - // optional bool is_guaranteed = 3; public static final int IS_GUARANTEED_FIELD_NUMBER = 3; private boolean isGuaranteed_; /** * optional bool is_guaranteed = 3; + * @return Whether the isGuaranteed field is set. */ + @java.lang.Override public boolean hasIsGuaranteed() { - return ((bitField0_ & 0x00000004) == 0x00000004); + return ((bitField0_ & 0x00000004) != 0); } /** * optional bool is_guaranteed = 3; + * @return The isGuaranteed. */ + @java.lang.Override public boolean getIsGuaranteed() { return isGuaranteed_; } - private void initFields() { - queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance(); - fragmentIdentifierString_ = ""; - isGuaranteed_ = false; - } private byte memoizedIsInitialized = -1; + @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } + @java.lang.Override public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); - if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeMessage(1, queryIdentifier_); + if (((bitField0_ & 0x00000001) != 0)) { + output.writeMessage(1, getQueryIdentifier()); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - output.writeBytes(2, getFragmentIdentifierStringBytes()); + if (((bitField0_ & 0x00000002) != 0)) { + com.google.protobuf.GeneratedMessageV3.writeString(output, 2, fragmentIdentifierString_); } - if (((bitField0_ & 0x00000004) == 0x00000004)) { + if (((bitField0_ & 0x00000004) != 0)) { output.writeBool(3, isGuaranteed_); } - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; + @java.lang.Override public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { + if (((bitField0_ & 0x00000001) != 0)) { size += com.google.protobuf.CodedOutputStream - .computeMessageSize(1, queryIdentifier_); + .computeMessageSize(1, getQueryIdentifier()); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(2, getFragmentIdentifierStringBytes()); + if (((bitField0_ & 0x00000002) != 0)) { + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(2, fragmentIdentifierString_); } - if (((bitField0_ & 0x00000004) == 0x00000004)) { + if (((bitField0_ & 0x00000004) != 0)) { size += com.google.protobuf.CodedOutputStream .computeBoolSize(3, isGuaranteed_); } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { @@ -16951,35 +19233,32 @@ public boolean equals(final java.lang.Object obj) { } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentRequestProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentRequestProto) obj; - boolean result = true; - result = result && (hasQueryIdentifier() == other.hasQueryIdentifier()); + if (hasQueryIdentifier() != other.hasQueryIdentifier()) return false; if (hasQueryIdentifier()) { - result = result && getQueryIdentifier() - .equals(other.getQueryIdentifier()); + if (!getQueryIdentifier() + .equals(other.getQueryIdentifier())) return false; } - result = result && (hasFragmentIdentifierString() == other.hasFragmentIdentifierString()); + if (hasFragmentIdentifierString() != other.hasFragmentIdentifierString()) return false; if (hasFragmentIdentifierString()) { - result = result && getFragmentIdentifierString() - .equals(other.getFragmentIdentifierString()); + if (!getFragmentIdentifierString() + .equals(other.getFragmentIdentifierString())) return false; } - result = result && (hasIsGuaranteed() == other.hasIsGuaranteed()); + if (hasIsGuaranteed() != other.hasIsGuaranteed()) return false; if (hasIsGuaranteed()) { - result = result && (getIsGuaranteed() - == other.getIsGuaranteed()); + if (getIsGuaranteed() + != other.getIsGuaranteed()) return false; } - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; + if (!unknownFields.equals(other.unknownFields)) return false; + return true; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); + hash = (19 * hash) + getDescriptor().hashCode(); if (hasQueryIdentifier()) { hash = (37 * hash) + QUERY_IDENTIFIER_FIELD_NUMBER; hash = (53 * hash) + getQueryIdentifier().hashCode(); @@ -16990,13 +19269,25 @@ public int hashCode() { } if (hasIsGuaranteed()) { hash = (37 * hash) + IS_GUARANTEED_FIELD_NUMBER; - hash = (53 * hash) + hashBoolean(getIsGuaranteed()); + hash = (53 * hash) + com.google.protobuf.Internal.hashBoolean( + getIsGuaranteed()); } - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentRequestProto parseFrom( + java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentRequestProto parseFrom( + java.nio.ByteBuffer data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentRequestProto parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { @@ -17020,46 +19311,59 @@ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.Up } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentRequestProto parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentRequestProto parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentRequestProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentRequestProto parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentRequestProto parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentRequestProto parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } + @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentRequestProto prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -17067,14 +19371,16 @@ protected Builder newBuilderForType( * Protobuf type {@code UpdateFragmentRequestProto} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentRequestProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder implements + // @@protoc_insertion_point(builder_implements:UpdateFragmentRequestProto) + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentRequestProtoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_UpdateFragmentRequestProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_UpdateFragmentRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -17087,23 +19393,21 @@ private Builder() { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { getQueryIdentifierFieldBuilder(); } } - private static Builder create() { - return new Builder(); - } - + @java.lang.Override public Builder clear() { super.clear(); if (queryIdentifierBuilder_ == null) { - queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance(); + queryIdentifier_ = null; } else { queryIdentifierBuilder_.clear(); } @@ -17115,19 +19419,18 @@ public Builder clear() { return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - + @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_UpdateFragmentRequestProto_descriptor; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentRequestProto getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentRequestProto.getDefaultInstance(); } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentRequestProto build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentRequestProto result = buildPartial(); if (!result.isInitialized()) { @@ -17136,31 +19439,65 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFra return result; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentRequestProto buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentRequestProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentRequestProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { + if (((from_bitField0_ & 0x00000001) != 0)) { + if (queryIdentifierBuilder_ == null) { + result.queryIdentifier_ = queryIdentifier_; + } else { + result.queryIdentifier_ = queryIdentifierBuilder_.build(); + } to_bitField0_ |= 0x00000001; } - if (queryIdentifierBuilder_ == null) { - result.queryIdentifier_ = queryIdentifier_; - } else { - result.queryIdentifier_ = queryIdentifierBuilder_.build(); - } - if (((from_bitField0_ & 0x00000002) == 0x00000002)) { + if (((from_bitField0_ & 0x00000002) != 0)) { to_bitField0_ |= 0x00000002; } result.fragmentIdentifierString_ = fragmentIdentifierString_; - if (((from_bitField0_ & 0x00000004) == 0x00000004)) { + if (((from_bitField0_ & 0x00000004) != 0)) { + result.isGuaranteed_ = isGuaranteed_; to_bitField0_ |= 0x00000004; } - result.isGuaranteed_ = isGuaranteed_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } + @java.lang.Override + public Builder clone() { + return super.clone(); + } + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.setField(field, value); + } + @java.lang.Override + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + @java.lang.Override + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.addRepeatedField(field, value); + } + @java.lang.Override public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentRequestProto) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentRequestProto)other); @@ -17183,14 +19520,17 @@ public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtoc if (other.hasIsGuaranteed()) { setIsGuaranteed(other.getIsGuaranteed()); } - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } + @java.lang.Override public final boolean isInitialized() { return true; } + @java.lang.Override public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) @@ -17200,7 +19540,7 @@ public Builder mergeFrom( parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentRequestProto) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -17210,22 +19550,23 @@ public Builder mergeFrom( } private int bitField0_; - // optional .QueryIdentifierProto query_identifier = 1; - private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance(); - private com.google.protobuf.SingleFieldBuilder< + private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto queryIdentifier_; + private com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder> queryIdentifierBuilder_; /** * optional .QueryIdentifierProto query_identifier = 1; + * @return Whether the queryIdentifier field is set. */ public boolean hasQueryIdentifier() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional .QueryIdentifierProto query_identifier = 1; + * @return The queryIdentifier. */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto getQueryIdentifier() { if (queryIdentifierBuilder_ == null) { - return queryIdentifier_; + return queryIdentifier_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance() : queryIdentifier_; } else { return queryIdentifierBuilder_.getMessage(); } @@ -17265,7 +19606,8 @@ public Builder setQueryIdentifier( */ public Builder mergeQueryIdentifier(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto value) { if (queryIdentifierBuilder_ == null) { - if (((bitField0_ & 0x00000001) == 0x00000001) && + if (((bitField0_ & 0x00000001) != 0) && + queryIdentifier_ != null && queryIdentifier_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance()) { queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.newBuilder(queryIdentifier_).mergeFrom(value).buildPartial(); @@ -17284,7 +19626,7 @@ public Builder mergeQueryIdentifier(org.apache.hadoop.hive.llap.daemon.rpc.LlapD */ public Builder clearQueryIdentifier() { if (queryIdentifierBuilder_ == null) { - queryIdentifier_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance(); + queryIdentifier_ = null; onChanged(); } else { queryIdentifierBuilder_.clear(); @@ -17307,19 +19649,20 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIden if (queryIdentifierBuilder_ != null) { return queryIdentifierBuilder_.getMessageOrBuilder(); } else { - return queryIdentifier_; + return queryIdentifier_ == null ? + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.getDefaultInstance() : queryIdentifier_; } } /** * optional .QueryIdentifierProto query_identifier = 1; */ - private com.google.protobuf.SingleFieldBuilder< + private com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder> getQueryIdentifierFieldBuilder() { if (queryIdentifierBuilder_ == null) { - queryIdentifierBuilder_ = new com.google.protobuf.SingleFieldBuilder< + queryIdentifierBuilder_ = new com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIdentifierProtoOrBuilder>( - queryIdentifier_, + getQueryIdentifier(), getParentForChildren(), isClean()); queryIdentifier_ = null; @@ -17327,23 +19670,27 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.QueryIden return queryIdentifierBuilder_; } - // optional string fragment_identifier_string = 2; private java.lang.Object fragmentIdentifierString_ = ""; /** * optional string fragment_identifier_string = 2; + * @return Whether the fragmentIdentifierString field is set. */ public boolean hasFragmentIdentifierString() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * optional string fragment_identifier_string = 2; + * @return The fragmentIdentifierString. */ public java.lang.String getFragmentIdentifierString() { java.lang.Object ref = fragmentIdentifierString_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - fragmentIdentifierString_ = s; + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + fragmentIdentifierString_ = s; + } return s; } else { return (java.lang.String) ref; @@ -17351,6 +19698,7 @@ public java.lang.String getFragmentIdentifierString() { } /** * optional string fragment_identifier_string = 2; + * @return The bytes for fragmentIdentifierString. */ public com.google.protobuf.ByteString getFragmentIdentifierStringBytes() { @@ -17367,6 +19715,8 @@ public java.lang.String getFragmentIdentifierString() { } /** * optional string fragment_identifier_string = 2; + * @param value The fragmentIdentifierString to set. + * @return This builder for chaining. */ public Builder setFragmentIdentifierString( java.lang.String value) { @@ -17380,6 +19730,7 @@ public Builder setFragmentIdentifierString( } /** * optional string fragment_identifier_string = 2; + * @return This builder for chaining. */ public Builder clearFragmentIdentifierString() { bitField0_ = (bitField0_ & ~0x00000002); @@ -17389,6 +19740,8 @@ public Builder clearFragmentIdentifierString() { } /** * optional string fragment_identifier_string = 2; + * @param value The bytes for fragmentIdentifierString to set. + * @return This builder for chaining. */ public Builder setFragmentIdentifierStringBytes( com.google.protobuf.ByteString value) { @@ -17401,22 +19754,27 @@ public Builder setFragmentIdentifierStringBytes( return this; } - // optional bool is_guaranteed = 3; private boolean isGuaranteed_ ; /** * optional bool is_guaranteed = 3; + * @return Whether the isGuaranteed field is set. */ + @java.lang.Override public boolean hasIsGuaranteed() { - return ((bitField0_ & 0x00000004) == 0x00000004); + return ((bitField0_ & 0x00000004) != 0); } /** * optional bool is_guaranteed = 3; + * @return The isGuaranteed. */ + @java.lang.Override public boolean getIsGuaranteed() { return isGuaranteed_; } /** * optional bool is_guaranteed = 3; + * @param value The isGuaranteed to set. + * @return This builder for chaining. */ public Builder setIsGuaranteed(boolean value) { bitField0_ |= 0x00000004; @@ -17426,6 +19784,7 @@ public Builder setIsGuaranteed(boolean value) { } /** * optional bool is_guaranteed = 3; + * @return This builder for chaining. */ public Builder clearIsGuaranteed() { bitField0_ = (bitField0_ & ~0x00000004); @@ -17433,38 +19792,82 @@ public Builder clearIsGuaranteed() { onChanged(); return this; } + @java.lang.Override + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:UpdateFragmentRequestProto) } + // @@protoc_insertion_point(class_scope:UpdateFragmentRequestProto) + private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentRequestProto DEFAULT_INSTANCE; static { - defaultInstance = new UpdateFragmentRequestProto(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentRequestProto(); + } + + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentRequestProto getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser + PARSER = new com.google.protobuf.AbstractParser() { + @java.lang.Override + public UpdateFragmentRequestProto parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new UpdateFragmentRequestProto(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentRequestProto getDefaultInstanceForType() { + return DEFAULT_INSTANCE; } - // @@protoc_insertion_point(class_scope:UpdateFragmentRequestProto) } - public interface UpdateFragmentResponseProtoOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface UpdateFragmentResponseProtoOrBuilder extends + // @@protoc_insertion_point(interface_extends:UpdateFragmentResponseProto) + com.google.protobuf.MessageOrBuilder { - // optional bool result = 1; /** * optional bool result = 1; + * @return Whether the result field is set. */ boolean hasResult(); /** * optional bool result = 1; + * @return The result. */ boolean getResult(); - // optional bool is_guaranteed = 2; /** * optional bool is_guaranteed = 2; + * @return Whether the isGuaranteed field is set. */ boolean hasIsGuaranteed(); /** * optional bool is_guaranteed = 2; + * @return The isGuaranteed. */ boolean getIsGuaranteed(); } @@ -17472,35 +19875,37 @@ public interface UpdateFragmentResponseProtoOrBuilder * Protobuf type {@code UpdateFragmentResponseProto} */ public static final class UpdateFragmentResponseProto extends - com.google.protobuf.GeneratedMessage - implements UpdateFragmentResponseProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:UpdateFragmentResponseProto) + UpdateFragmentResponseProtoOrBuilder { + private static final long serialVersionUID = 0L; // Use UpdateFragmentResponseProto.newBuilder() to construct. - private UpdateFragmentResponseProto(com.google.protobuf.GeneratedMessage.Builder builder) { + private UpdateFragmentResponseProto(com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private UpdateFragmentResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final UpdateFragmentResponseProto defaultInstance; - public static UpdateFragmentResponseProto getDefaultInstance() { - return defaultInstance; + private UpdateFragmentResponseProto() { } - public UpdateFragmentResponseProto getDefaultInstanceForType() { - return defaultInstance; + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance( + UnusedPrivateParameter unused) { + return new UpdateFragmentResponseProto(); } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private UpdateFragmentResponseProto( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); @@ -17512,13 +19917,6 @@ private UpdateFragmentResponseProto( case 0: done = true; break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } case 8: { bitField0_ |= 0x00000001; result_ = input.readBool(); @@ -17529,13 +19927,20 @@ private UpdateFragmentResponseProto( isGuaranteed_ = input.readBool(); break; } + default: { + if (!parseUnknownField( + input, unknownFields, extensionRegistry, tag)) { + done = true; + } + break; + } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); @@ -17546,112 +19951,95 @@ private UpdateFragmentResponseProto( return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_UpdateFragmentResponseProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_UpdateFragmentResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentResponseProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentResponseProto.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public UpdateFragmentResponseProto parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new UpdateFragmentResponseProto(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - private int bitField0_; - // optional bool result = 1; public static final int RESULT_FIELD_NUMBER = 1; private boolean result_; /** * optional bool result = 1; + * @return Whether the result field is set. */ + @java.lang.Override public boolean hasResult() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional bool result = 1; + * @return The result. */ + @java.lang.Override public boolean getResult() { return result_; } - // optional bool is_guaranteed = 2; public static final int IS_GUARANTEED_FIELD_NUMBER = 2; private boolean isGuaranteed_; /** * optional bool is_guaranteed = 2; + * @return Whether the isGuaranteed field is set. */ + @java.lang.Override public boolean hasIsGuaranteed() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * optional bool is_guaranteed = 2; + * @return The isGuaranteed. */ + @java.lang.Override public boolean getIsGuaranteed() { return isGuaranteed_; } - private void initFields() { - result_ = false; - isGuaranteed_ = false; - } private byte memoizedIsInitialized = -1; + @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } + @java.lang.Override public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); - if (((bitField0_ & 0x00000001) == 0x00000001)) { + if (((bitField0_ & 0x00000001) != 0)) { output.writeBool(1, result_); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { + if (((bitField0_ & 0x00000002) != 0)) { output.writeBool(2, isGuaranteed_); } - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; + @java.lang.Override public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { + if (((bitField0_ & 0x00000001) != 0)) { size += com.google.protobuf.CodedOutputStream .computeBoolSize(1, result_); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { + if (((bitField0_ & 0x00000002) != 0)) { size += com.google.protobuf.CodedOutputStream .computeBoolSize(2, isGuaranteed_); } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { @@ -17662,43 +20050,53 @@ public boolean equals(final java.lang.Object obj) { } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentResponseProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentResponseProto) obj; - boolean result = true; - result = result && (hasResult() == other.hasResult()); + if (hasResult() != other.hasResult()) return false; if (hasResult()) { - result = result && (getResult() - == other.getResult()); + if (getResult() + != other.getResult()) return false; } - result = result && (hasIsGuaranteed() == other.hasIsGuaranteed()); + if (hasIsGuaranteed() != other.hasIsGuaranteed()) return false; if (hasIsGuaranteed()) { - result = result && (getIsGuaranteed() - == other.getIsGuaranteed()); + if (getIsGuaranteed() + != other.getIsGuaranteed()) return false; } - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; + if (!unknownFields.equals(other.unknownFields)) return false; + return true; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); + hash = (19 * hash) + getDescriptor().hashCode(); if (hasResult()) { hash = (37 * hash) + RESULT_FIELD_NUMBER; - hash = (53 * hash) + hashBoolean(getResult()); + hash = (53 * hash) + com.google.protobuf.Internal.hashBoolean( + getResult()); } if (hasIsGuaranteed()) { hash = (37 * hash) + IS_GUARANTEED_FIELD_NUMBER; - hash = (53 * hash) + hashBoolean(getIsGuaranteed()); + hash = (53 * hash) + com.google.protobuf.Internal.hashBoolean( + getIsGuaranteed()); } - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentResponseProto parseFrom( + java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentResponseProto parseFrom( + java.nio.ByteBuffer data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentResponseProto parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { @@ -17722,46 +20120,59 @@ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.Up } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentResponseProto parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentResponseProto parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentResponseProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentResponseProto parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentResponseProto parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentResponseProto parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } + @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentResponseProto prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -17769,14 +20180,16 @@ protected Builder newBuilderForType( * Protobuf type {@code UpdateFragmentResponseProto} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentResponseProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder implements + // @@protoc_insertion_point(builder_implements:UpdateFragmentResponseProto) + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentResponseProtoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_UpdateFragmentResponseProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_UpdateFragmentResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -17789,18 +20202,16 @@ private Builder() { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { } } - private static Builder create() { - return new Builder(); - } - + @java.lang.Override public Builder clear() { super.clear(); result_ = false; @@ -17810,19 +20221,18 @@ public Builder clear() { return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - + @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_UpdateFragmentResponseProto_descriptor; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentResponseProto getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentResponseProto.getDefaultInstance(); } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentResponseProto build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentResponseProto result = buildPartial(); if (!result.isInitialized()) { @@ -17831,23 +20241,57 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFra return result; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentResponseProto buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentResponseProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentResponseProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { + if (((from_bitField0_ & 0x00000001) != 0)) { + result.result_ = result_; to_bitField0_ |= 0x00000001; } - result.result_ = result_; - if (((from_bitField0_ & 0x00000002) == 0x00000002)) { + if (((from_bitField0_ & 0x00000002) != 0)) { + result.isGuaranteed_ = isGuaranteed_; to_bitField0_ |= 0x00000002; } - result.isGuaranteed_ = isGuaranteed_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } + @java.lang.Override + public Builder clone() { + return super.clone(); + } + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.setField(field, value); + } + @java.lang.Override + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + @java.lang.Override + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.addRepeatedField(field, value); + } + @java.lang.Override public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentResponseProto) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentResponseProto)other); @@ -17865,14 +20309,17 @@ public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtoc if (other.hasIsGuaranteed()) { setIsGuaranteed(other.getIsGuaranteed()); } - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } + @java.lang.Override public final boolean isInitialized() { return true; } + @java.lang.Override public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) @@ -17882,7 +20329,7 @@ public Builder mergeFrom( parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentResponseProto) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -17892,22 +20339,27 @@ public Builder mergeFrom( } private int bitField0_; - // optional bool result = 1; private boolean result_ ; /** * optional bool result = 1; + * @return Whether the result field is set. */ + @java.lang.Override public boolean hasResult() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional bool result = 1; + * @return The result. */ + @java.lang.Override public boolean getResult() { return result_; } /** * optional bool result = 1; + * @param value The result to set. + * @return This builder for chaining. */ public Builder setResult(boolean value) { bitField0_ |= 0x00000001; @@ -17917,6 +20369,7 @@ public Builder setResult(boolean value) { } /** * optional bool result = 1; + * @return This builder for chaining. */ public Builder clearResult() { bitField0_ = (bitField0_ & ~0x00000001); @@ -17925,22 +20378,27 @@ public Builder clearResult() { return this; } - // optional bool is_guaranteed = 2; private boolean isGuaranteed_ ; /** * optional bool is_guaranteed = 2; + * @return Whether the isGuaranteed field is set. */ + @java.lang.Override public boolean hasIsGuaranteed() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * optional bool is_guaranteed = 2; + * @return The isGuaranteed. */ + @java.lang.Override public boolean getIsGuaranteed() { return isGuaranteed_; } /** * optional bool is_guaranteed = 2; + * @param value The isGuaranteed to set. + * @return This builder for chaining. */ public Builder setIsGuaranteed(boolean value) { bitField0_ |= 0x00000002; @@ -17950,6 +20408,7 @@ public Builder setIsGuaranteed(boolean value) { } /** * optional bool is_guaranteed = 2; + * @return This builder for chaining. */ public Builder clearIsGuaranteed() { bitField0_ = (bitField0_ & ~0x00000002); @@ -17957,32 +20416,76 @@ public Builder clearIsGuaranteed() { onChanged(); return this; } + @java.lang.Override + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:UpdateFragmentResponseProto) } + // @@protoc_insertion_point(class_scope:UpdateFragmentResponseProto) + private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentResponseProto DEFAULT_INSTANCE; static { - defaultInstance = new UpdateFragmentResponseProto(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentResponseProto(); + } + + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentResponseProto getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser + PARSER = new com.google.protobuf.AbstractParser() { + @java.lang.Override + public UpdateFragmentResponseProto parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new UpdateFragmentResponseProto(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.UpdateFragmentResponseProto getDefaultInstanceForType() { + return DEFAULT_INSTANCE; } - // @@protoc_insertion_point(class_scope:UpdateFragmentResponseProto) } - public interface GetTokenRequestProtoOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface GetTokenRequestProtoOrBuilder extends + // @@protoc_insertion_point(interface_extends:GetTokenRequestProto) + com.google.protobuf.MessageOrBuilder { - // optional string app_id = 1; /** * optional string app_id = 1; + * @return Whether the appId field is set. */ boolean hasAppId(); /** * optional string app_id = 1; + * @return The appId. */ java.lang.String getAppId(); /** * optional string app_id = 1; + * @return The bytes for appId. */ com.google.protobuf.ByteString getAppIdBytes(); @@ -17991,35 +20494,38 @@ public interface GetTokenRequestProtoOrBuilder * Protobuf type {@code GetTokenRequestProto} */ public static final class GetTokenRequestProto extends - com.google.protobuf.GeneratedMessage - implements GetTokenRequestProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:GetTokenRequestProto) + GetTokenRequestProtoOrBuilder { + private static final long serialVersionUID = 0L; // Use GetTokenRequestProto.newBuilder() to construct. - private GetTokenRequestProto(com.google.protobuf.GeneratedMessage.Builder builder) { + private GetTokenRequestProto(com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private GetTokenRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final GetTokenRequestProto defaultInstance; - public static GetTokenRequestProto getDefaultInstance() { - return defaultInstance; + private GetTokenRequestProto() { + appId_ = ""; } - public GetTokenRequestProto getDefaultInstanceForType() { - return defaultInstance; + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance( + UnusedPrivateParameter unused) { + return new GetTokenRequestProto(); } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private GetTokenRequestProto( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); @@ -18031,25 +20537,26 @@ private GetTokenRequestProto( case 0: done = true; break; + case 10: { + com.google.protobuf.ByteString bs = input.readBytes(); + bitField0_ |= 0x00000001; + appId_ = bs; + break; + } default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { + if (!parseUnknownField( + input, unknownFields, extensionRegistry, tag)) { done = true; } break; } - case 10: { - bitField0_ |= 0x00000001; - appId_ = input.readBytes(); - break; - } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); @@ -18060,41 +20567,30 @@ private GetTokenRequestProto( return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GetTokenRequestProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GetTokenRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public GetTokenRequestProto parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new GetTokenRequestProto(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - private int bitField0_; - // optional string app_id = 1; public static final int APP_ID_FIELD_NUMBER = 1; - private java.lang.Object appId_; + private volatile java.lang.Object appId_; /** * optional string app_id = 1; + * @return Whether the appId field is set. */ + @java.lang.Override public boolean hasAppId() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional string app_id = 1; + * @return The appId. */ + @java.lang.Override public java.lang.String getAppId() { java.lang.Object ref = appId_; if (ref instanceof java.lang.String) { @@ -18111,7 +20607,9 @@ public java.lang.String getAppId() { } /** * optional string app_id = 1; + * @return The bytes for appId. */ + @java.lang.Override public com.google.protobuf.ByteString getAppIdBytes() { java.lang.Object ref = appId_; @@ -18126,49 +20624,40 @@ public java.lang.String getAppId() { } } - private void initFields() { - appId_ = ""; - } private byte memoizedIsInitialized = -1; + @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } + @java.lang.Override public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); - if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeBytes(1, getAppIdBytes()); + if (((bitField0_ & 0x00000001) != 0)) { + com.google.protobuf.GeneratedMessageV3.writeString(output, 1, appId_); } - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; + @java.lang.Override public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(1, getAppIdBytes()); + if (((bitField0_ & 0x00000001) != 0)) { + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, appId_); } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { @@ -18179,34 +20668,42 @@ public boolean equals(final java.lang.Object obj) { } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto) obj; - boolean result = true; - result = result && (hasAppId() == other.hasAppId()); + if (hasAppId() != other.hasAppId()) return false; if (hasAppId()) { - result = result && getAppId() - .equals(other.getAppId()); + if (!getAppId() + .equals(other.getAppId())) return false; } - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; + if (!unknownFields.equals(other.unknownFields)) return false; + return true; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); + hash = (19 * hash) + getDescriptor().hashCode(); if (hasAppId()) { hash = (37 * hash) + APP_ID_FIELD_NUMBER; hash = (53 * hash) + getAppId().hashCode(); } - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto parseFrom( + java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto parseFrom( + java.nio.ByteBuffer data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { @@ -18230,46 +20727,59 @@ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.Ge } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } + @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -18277,14 +20787,16 @@ protected Builder newBuilderForType( * Protobuf type {@code GetTokenRequestProto} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder implements + // @@protoc_insertion_point(builder_implements:GetTokenRequestProto) + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProtoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GetTokenRequestProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GetTokenRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -18297,18 +20809,16 @@ private Builder() { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { } } - private static Builder create() { - return new Builder(); - } - + @java.lang.Override public Builder clear() { super.clear(); appId_ = ""; @@ -18316,19 +20826,18 @@ public Builder clear() { return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - + @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GetTokenRequestProto_descriptor; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto.getDefaultInstance(); } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto result = buildPartial(); if (!result.isInitialized()) { @@ -18337,11 +20846,12 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenR return result; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { + if (((from_bitField0_ & 0x00000001) != 0)) { to_bitField0_ |= 0x00000001; } result.appId_ = appId_; @@ -18350,6 +20860,39 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenR return result; } + @java.lang.Override + public Builder clone() { + return super.clone(); + } + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.setField(field, value); + } + @java.lang.Override + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + @java.lang.Override + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.addRepeatedField(field, value); + } + @java.lang.Override public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto)other); @@ -18366,14 +20909,17 @@ public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtoc appId_ = other.appId_; onChanged(); } - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } + @java.lang.Override public final boolean isInitialized() { return true; } + @java.lang.Override public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) @@ -18383,7 +20929,7 @@ public Builder mergeFrom( parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -18393,23 +20939,27 @@ public Builder mergeFrom( } private int bitField0_; - // optional string app_id = 1; private java.lang.Object appId_ = ""; /** * optional string app_id = 1; + * @return Whether the appId field is set. */ public boolean hasAppId() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional string app_id = 1; + * @return The appId. */ public java.lang.String getAppId() { java.lang.Object ref = appId_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - appId_ = s; + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + appId_ = s; + } return s; } else { return (java.lang.String) ref; @@ -18417,6 +20967,7 @@ public java.lang.String getAppId() { } /** * optional string app_id = 1; + * @return The bytes for appId. */ public com.google.protobuf.ByteString getAppIdBytes() { @@ -18433,6 +20984,8 @@ public java.lang.String getAppId() { } /** * optional string app_id = 1; + * @param value The appId to set. + * @return This builder for chaining. */ public Builder setAppId( java.lang.String value) { @@ -18446,6 +20999,7 @@ public Builder setAppId( } /** * optional string app_id = 1; + * @return This builder for chaining. */ public Builder clearAppId() { bitField0_ = (bitField0_ & ~0x00000001); @@ -18455,6 +21009,8 @@ public Builder clearAppId() { } /** * optional string app_id = 1; + * @param value The bytes for appId to set. + * @return This builder for chaining. */ public Builder setAppIdBytes( com.google.protobuf.ByteString value) { @@ -18466,28 +21022,71 @@ public Builder setAppIdBytes( onChanged(); return this; } + @java.lang.Override + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:GetTokenRequestProto) } + // @@protoc_insertion_point(class_scope:GetTokenRequestProto) + private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto DEFAULT_INSTANCE; static { - defaultInstance = new GetTokenRequestProto(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto(); + } + + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser + PARSER = new com.google.protobuf.AbstractParser() { + @java.lang.Override + public GetTokenRequestProto parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new GetTokenRequestProto(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenRequestProto getDefaultInstanceForType() { + return DEFAULT_INSTANCE; } - // @@protoc_insertion_point(class_scope:GetTokenRequestProto) } - public interface GetTokenResponseProtoOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface GetTokenResponseProtoOrBuilder extends + // @@protoc_insertion_point(interface_extends:GetTokenResponseProto) + com.google.protobuf.MessageOrBuilder { - // optional bytes token = 1; /** * optional bytes token = 1; + * @return Whether the token field is set. */ boolean hasToken(); /** * optional bytes token = 1; + * @return The token. */ com.google.protobuf.ByteString getToken(); } @@ -18495,35 +21094,38 @@ public interface GetTokenResponseProtoOrBuilder * Protobuf type {@code GetTokenResponseProto} */ public static final class GetTokenResponseProto extends - com.google.protobuf.GeneratedMessage - implements GetTokenResponseProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:GetTokenResponseProto) + GetTokenResponseProtoOrBuilder { + private static final long serialVersionUID = 0L; // Use GetTokenResponseProto.newBuilder() to construct. - private GetTokenResponseProto(com.google.protobuf.GeneratedMessage.Builder builder) { + private GetTokenResponseProto(com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private GetTokenResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final GetTokenResponseProto defaultInstance; - public static GetTokenResponseProto getDefaultInstance() { - return defaultInstance; + private GetTokenResponseProto() { + token_ = com.google.protobuf.ByteString.EMPTY; } - public GetTokenResponseProto getDefaultInstanceForType() { - return defaultInstance; + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance( + UnusedPrivateParameter unused) { + return new GetTokenResponseProto(); } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private GetTokenResponseProto( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); @@ -18535,25 +21137,25 @@ private GetTokenResponseProto( case 0: done = true; break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } case 10: { bitField0_ |= 0x00000001; token_ = input.readBytes(); break; } + default: { + if (!parseUnknownField( + input, unknownFields, extensionRegistry, tag)) { + done = true; + } + break; + } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); @@ -18564,88 +21166,69 @@ private GetTokenResponseProto( return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GetTokenResponseProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GetTokenResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public GetTokenResponseProto parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new GetTokenResponseProto(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - private int bitField0_; - // optional bytes token = 1; public static final int TOKEN_FIELD_NUMBER = 1; private com.google.protobuf.ByteString token_; /** * optional bytes token = 1; + * @return Whether the token field is set. */ + @java.lang.Override public boolean hasToken() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional bytes token = 1; + * @return The token. */ + @java.lang.Override public com.google.protobuf.ByteString getToken() { return token_; } - private void initFields() { - token_ = com.google.protobuf.ByteString.EMPTY; - } private byte memoizedIsInitialized = -1; + @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } + @java.lang.Override public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); - if (((bitField0_ & 0x00000001) == 0x00000001)) { + if (((bitField0_ & 0x00000001) != 0)) { output.writeBytes(1, token_); } - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; + @java.lang.Override public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { + if (((bitField0_ & 0x00000001) != 0)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(1, token_); } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { @@ -18656,41 +21239,49 @@ public boolean equals(final java.lang.Object obj) { } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto) obj; - boolean result = true; - result = result && (hasToken() == other.hasToken()); + if (hasToken() != other.hasToken()) return false; if (hasToken()) { - result = result && getToken() - .equals(other.getToken()); + if (!getToken() + .equals(other.getToken())) return false; } - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; + if (!unknownFields.equals(other.unknownFields)) return false; + return true; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); + hash = (19 * hash) + getDescriptor().hashCode(); if (hasToken()) { hash = (37 * hash) + TOKEN_FIELD_NUMBER; hash = (53 * hash) + getToken().hashCode(); } - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto parseFrom( - com.google.protobuf.ByteString data) + java.nio.ByteBuffer data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto parseFrom( - com.google.protobuf.ByteString data, + java.nio.ByteBuffer data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto parseFrom( + com.google.protobuf.ByteString data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto parseFrom( + com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); @@ -18707,46 +21298,59 @@ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.Ge } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } + @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -18754,14 +21358,16 @@ protected Builder newBuilderForType( * Protobuf type {@code GetTokenResponseProto} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder implements + // @@protoc_insertion_point(builder_implements:GetTokenResponseProto) + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProtoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GetTokenResponseProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GetTokenResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -18774,18 +21380,16 @@ private Builder() { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { } } - private static Builder create() { - return new Builder(); - } - + @java.lang.Override public Builder clear() { super.clear(); token_ = com.google.protobuf.ByteString.EMPTY; @@ -18793,19 +21397,18 @@ public Builder clear() { return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - + @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GetTokenResponseProto_descriptor; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto.getDefaultInstance(); } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto result = buildPartial(); if (!result.isInitialized()) { @@ -18814,11 +21417,12 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenR return result; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { + if (((from_bitField0_ & 0x00000001) != 0)) { to_bitField0_ |= 0x00000001; } result.token_ = token_; @@ -18827,6 +21431,39 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenR return result; } + @java.lang.Override + public Builder clone() { + return super.clone(); + } + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.setField(field, value); + } + @java.lang.Override + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + @java.lang.Override + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.addRepeatedField(field, value); + } + @java.lang.Override public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto)other); @@ -18841,14 +21478,17 @@ public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtoc if (other.hasToken()) { setToken(other.getToken()); } - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } + @java.lang.Override public final boolean isInitialized() { return true; } + @java.lang.Override public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) @@ -18858,7 +21498,7 @@ public Builder mergeFrom( parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -18868,22 +21508,27 @@ public Builder mergeFrom( } private int bitField0_; - // optional bytes token = 1; private com.google.protobuf.ByteString token_ = com.google.protobuf.ByteString.EMPTY; /** * optional bytes token = 1; + * @return Whether the token field is set. */ + @java.lang.Override public boolean hasToken() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional bytes token = 1; + * @return The token. */ + @java.lang.Override public com.google.protobuf.ByteString getToken() { return token_; } /** * optional bytes token = 1; + * @param value The token to set. + * @return This builder for chaining. */ public Builder setToken(com.google.protobuf.ByteString value) { if (value == null) { @@ -18896,6 +21541,7 @@ public Builder setToken(com.google.protobuf.ByteString value) { } /** * optional bytes token = 1; + * @return This builder for chaining. */ public Builder clearToken() { bitField0_ = (bitField0_ & ~0x00000001); @@ -18903,83 +21549,132 @@ public Builder clearToken() { onChanged(); return this; } + @java.lang.Override + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:GetTokenResponseProto) } + // @@protoc_insertion_point(class_scope:GetTokenResponseProto) + private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto DEFAULT_INSTANCE; static { - defaultInstance = new GetTokenResponseProto(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto(); + } + + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser + PARSER = new com.google.protobuf.AbstractParser() { + @java.lang.Override + public GetTokenResponseProto parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new GetTokenResponseProto(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetTokenResponseProto getDefaultInstanceForType() { + return DEFAULT_INSTANCE; } - // @@protoc_insertion_point(class_scope:GetTokenResponseProto) } - public interface LlapOutputSocketInitMessageOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface LlapOutputSocketInitMessageOrBuilder extends + // @@protoc_insertion_point(interface_extends:LlapOutputSocketInitMessage) + com.google.protobuf.MessageOrBuilder { - // required string fragment_id = 1; /** * required string fragment_id = 1; + * @return Whether the fragmentId field is set. */ boolean hasFragmentId(); /** * required string fragment_id = 1; + * @return The fragmentId. */ java.lang.String getFragmentId(); /** * required string fragment_id = 1; + * @return The bytes for fragmentId. */ com.google.protobuf.ByteString getFragmentIdBytes(); - // optional bytes token = 2; /** * optional bytes token = 2; + * @return Whether the token field is set. */ boolean hasToken(); /** * optional bytes token = 2; + * @return The token. */ com.google.protobuf.ByteString getToken(); } /** - * Protobuf type {@code LlapOutputSocketInitMessage} - * *
    * The message sent by external client to claim the output from the output socket.
    * 
+ * + * Protobuf type {@code LlapOutputSocketInitMessage} */ public static final class LlapOutputSocketInitMessage extends - com.google.protobuf.GeneratedMessage - implements LlapOutputSocketInitMessageOrBuilder { + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:LlapOutputSocketInitMessage) + LlapOutputSocketInitMessageOrBuilder { + private static final long serialVersionUID = 0L; // Use LlapOutputSocketInitMessage.newBuilder() to construct. - private LlapOutputSocketInitMessage(com.google.protobuf.GeneratedMessage.Builder builder) { + private LlapOutputSocketInitMessage(com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private LlapOutputSocketInitMessage(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final LlapOutputSocketInitMessage defaultInstance; - public static LlapOutputSocketInitMessage getDefaultInstance() { - return defaultInstance; + private LlapOutputSocketInitMessage() { + fragmentId_ = ""; + token_ = com.google.protobuf.ByteString.EMPTY; } - public LlapOutputSocketInitMessage getDefaultInstanceForType() { - return defaultInstance; + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance( + UnusedPrivateParameter unused) { + return new LlapOutputSocketInitMessage(); } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private LlapOutputSocketInitMessage( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); @@ -18991,16 +21686,10 @@ private LlapOutputSocketInitMessage( case 0: done = true; break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } case 10: { + com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000001; - fragmentId_ = input.readBytes(); + fragmentId_ = bs; break; } case 18: { @@ -19008,13 +21697,20 @@ private LlapOutputSocketInitMessage( token_ = input.readBytes(); break; } + default: { + if (!parseUnknownField( + input, unknownFields, extensionRegistry, tag)) { + done = true; + } + break; + } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); @@ -19025,41 +21721,30 @@ private LlapOutputSocketInitMessage( return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_LlapOutputSocketInitMessage_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_LlapOutputSocketInitMessage_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public LlapOutputSocketInitMessage parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new LlapOutputSocketInitMessage(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - private int bitField0_; - // required string fragment_id = 1; public static final int FRAGMENT_ID_FIELD_NUMBER = 1; - private java.lang.Object fragmentId_; + private volatile java.lang.Object fragmentId_; /** * required string fragment_id = 1; + * @return Whether the fragmentId field is set. */ + @java.lang.Override public boolean hasFragmentId() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * required string fragment_id = 1; + * @return The fragmentId. */ + @java.lang.Override public java.lang.String getFragmentId() { java.lang.Object ref = fragmentId_; if (ref instanceof java.lang.String) { @@ -19076,7 +21761,9 @@ public java.lang.String getFragmentId() { } /** * required string fragment_id = 1; + * @return The bytes for fragmentId. */ + @java.lang.Override public com.google.protobuf.ByteString getFragmentIdBytes() { java.lang.Object ref = fragmentId_; @@ -19091,30 +21778,31 @@ public java.lang.String getFragmentId() { } } - // optional bytes token = 2; public static final int TOKEN_FIELD_NUMBER = 2; private com.google.protobuf.ByteString token_; /** * optional bytes token = 2; + * @return Whether the token field is set. */ + @java.lang.Override public boolean hasToken() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * optional bytes token = 2; + * @return The token. */ + @java.lang.Override public com.google.protobuf.ByteString getToken() { return token_; } - private void initFields() { - fragmentId_ = ""; - token_ = com.google.protobuf.ByteString.EMPTY; - } private byte memoizedIsInitialized = -1; + @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; if (!hasFragmentId()) { memoizedIsInitialized = 0; @@ -19124,44 +21812,36 @@ public final boolean isInitialized() { return true; } + @java.lang.Override public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); - if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeBytes(1, getFragmentIdBytes()); + if (((bitField0_ & 0x00000001) != 0)) { + com.google.protobuf.GeneratedMessageV3.writeString(output, 1, fragmentId_); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { + if (((bitField0_ & 0x00000002) != 0)) { output.writeBytes(2, token_); } - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; + @java.lang.Override public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(1, getFragmentIdBytes()); + if (((bitField0_ & 0x00000001) != 0)) { + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, fragmentId_); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { + if (((bitField0_ & 0x00000002) != 0)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(2, token_); } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { @@ -19172,30 +21852,27 @@ public boolean equals(final java.lang.Object obj) { } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage) obj; - boolean result = true; - result = result && (hasFragmentId() == other.hasFragmentId()); + if (hasFragmentId() != other.hasFragmentId()) return false; if (hasFragmentId()) { - result = result && getFragmentId() - .equals(other.getFragmentId()); + if (!getFragmentId() + .equals(other.getFragmentId())) return false; } - result = result && (hasToken() == other.hasToken()); + if (hasToken() != other.hasToken()) return false; if (hasToken()) { - result = result && getToken() - .equals(other.getToken()); + if (!getToken() + .equals(other.getToken())) return false; } - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; + if (!unknownFields.equals(other.unknownFields)) return false; + return true; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); + hash = (19 * hash) + getDescriptor().hashCode(); if (hasFragmentId()) { hash = (37 * hash) + FRAGMENT_ID_FIELD_NUMBER; hash = (53 * hash) + getFragmentId().hashCode(); @@ -19204,11 +21881,22 @@ public int hashCode() { hash = (37 * hash) + TOKEN_FIELD_NUMBER; hash = (53 * hash) + getToken().hashCode(); } - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage parseFrom( + java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage parseFrom( + java.nio.ByteBuffer data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { @@ -19232,65 +21920,80 @@ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.Ll } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } + @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** - * Protobuf type {@code LlapOutputSocketInitMessage} - * *
      * The message sent by external client to claim the output from the output socket.
      * 
+ * + * Protobuf type {@code LlapOutputSocketInitMessage} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessageOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder implements + // @@protoc_insertion_point(builder_implements:LlapOutputSocketInitMessage) + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessageOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_LlapOutputSocketInitMessage_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_LlapOutputSocketInitMessage_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -19303,18 +22006,16 @@ private Builder() { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { } } - private static Builder create() { - return new Builder(); - } - + @java.lang.Override public Builder clear() { super.clear(); fragmentId_ = ""; @@ -19324,19 +22025,18 @@ public Builder clear() { return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - + @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_LlapOutputSocketInitMessage_descriptor; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage.getDefaultInstance(); } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage result = buildPartial(); if (!result.isInitialized()) { @@ -19345,15 +22045,16 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutpu return result; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { + if (((from_bitField0_ & 0x00000001) != 0)) { to_bitField0_ |= 0x00000001; } result.fragmentId_ = fragmentId_; - if (((from_bitField0_ & 0x00000002) == 0x00000002)) { + if (((from_bitField0_ & 0x00000002) != 0)) { to_bitField0_ |= 0x00000002; } result.token_ = token_; @@ -19362,6 +22063,39 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutpu return result; } + @java.lang.Override + public Builder clone() { + return super.clone(); + } + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.setField(field, value); + } + @java.lang.Override + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + @java.lang.Override + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.addRepeatedField(field, value); + } + @java.lang.Override public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage)other); @@ -19381,18 +22115,20 @@ public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtoc if (other.hasToken()) { setToken(other.getToken()); } - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } + @java.lang.Override public final boolean isInitialized() { if (!hasFragmentId()) { - return false; } return true; } + @java.lang.Override public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) @@ -19402,7 +22138,7 @@ public Builder mergeFrom( parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -19412,23 +22148,27 @@ public Builder mergeFrom( } private int bitField0_; - // required string fragment_id = 1; private java.lang.Object fragmentId_ = ""; /** * required string fragment_id = 1; + * @return Whether the fragmentId field is set. */ public boolean hasFragmentId() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * required string fragment_id = 1; + * @return The fragmentId. */ public java.lang.String getFragmentId() { java.lang.Object ref = fragmentId_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - fragmentId_ = s; + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + fragmentId_ = s; + } return s; } else { return (java.lang.String) ref; @@ -19436,6 +22176,7 @@ public java.lang.String getFragmentId() { } /** * required string fragment_id = 1; + * @return The bytes for fragmentId. */ public com.google.protobuf.ByteString getFragmentIdBytes() { @@ -19452,6 +22193,8 @@ public java.lang.String getFragmentId() { } /** * required string fragment_id = 1; + * @param value The fragmentId to set. + * @return This builder for chaining. */ public Builder setFragmentId( java.lang.String value) { @@ -19465,6 +22208,7 @@ public Builder setFragmentId( } /** * required string fragment_id = 1; + * @return This builder for chaining. */ public Builder clearFragmentId() { bitField0_ = (bitField0_ & ~0x00000001); @@ -19474,6 +22218,8 @@ public Builder clearFragmentId() { } /** * required string fragment_id = 1; + * @param value The bytes for fragmentId to set. + * @return This builder for chaining. */ public Builder setFragmentIdBytes( com.google.protobuf.ByteString value) { @@ -19486,22 +22232,27 @@ public Builder setFragmentIdBytes( return this; } - // optional bytes token = 2; private com.google.protobuf.ByteString token_ = com.google.protobuf.ByteString.EMPTY; /** * optional bytes token = 2; + * @return Whether the token field is set. */ + @java.lang.Override public boolean hasToken() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * optional bytes token = 2; + * @return The token. */ + @java.lang.Override public com.google.protobuf.ByteString getToken() { return token_; } /** * optional bytes token = 2; + * @param value The token to set. + * @return This builder for chaining. */ public Builder setToken(com.google.protobuf.ByteString value) { if (value == null) { @@ -19514,6 +22265,7 @@ public Builder setToken(com.google.protobuf.ByteString value) { } /** * optional bytes token = 2; + * @return This builder for chaining. */ public Builder clearToken() { bitField0_ = (bitField0_ & ~0x00000002); @@ -19521,54 +22273,98 @@ public Builder clearToken() { onChanged(); return this; } + @java.lang.Override + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:LlapOutputSocketInitMessage) } + // @@protoc_insertion_point(class_scope:LlapOutputSocketInitMessage) + private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage DEFAULT_INSTANCE; static { - defaultInstance = new LlapOutputSocketInitMessage(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage(); + } + + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser + PARSER = new com.google.protobuf.AbstractParser() { + @java.lang.Override + public LlapOutputSocketInitMessage parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new LlapOutputSocketInitMessage(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.LlapOutputSocketInitMessage getDefaultInstanceForType() { + return DEFAULT_INSTANCE; } - // @@protoc_insertion_point(class_scope:LlapOutputSocketInitMessage) } - public interface PurgeCacheRequestProtoOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface PurgeCacheRequestProtoOrBuilder extends + // @@protoc_insertion_point(interface_extends:PurgeCacheRequestProto) + com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code PurgeCacheRequestProto} */ public static final class PurgeCacheRequestProto extends - com.google.protobuf.GeneratedMessage - implements PurgeCacheRequestProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:PurgeCacheRequestProto) + PurgeCacheRequestProtoOrBuilder { + private static final long serialVersionUID = 0L; // Use PurgeCacheRequestProto.newBuilder() to construct. - private PurgeCacheRequestProto(com.google.protobuf.GeneratedMessage.Builder builder) { + private PurgeCacheRequestProto(com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private PurgeCacheRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final PurgeCacheRequestProto defaultInstance; - public static PurgeCacheRequestProto getDefaultInstance() { - return defaultInstance; + private PurgeCacheRequestProto() { } - public PurgeCacheRequestProto getDefaultInstanceForType() { - return defaultInstance; + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance( + UnusedPrivateParameter unused) { + return new PurgeCacheRequestProto(); } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private PurgeCacheRequestProto( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); try { @@ -19580,8 +22376,8 @@ private PurgeCacheRequestProto( done = true; break; default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { + if (!parseUnknownField( + input, unknownFields, extensionRegistry, tag)) { done = true; } break; @@ -19592,7 +22388,7 @@ private PurgeCacheRequestProto( throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); @@ -19603,63 +22399,42 @@ private PurgeCacheRequestProto( return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_PurgeCacheRequestProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_PurgeCacheRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheRequestProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheRequestProto.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public PurgeCacheRequestProto parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new PurgeCacheRequestProto(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - - private void initFields() { - } private byte memoizedIsInitialized = -1; + @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } + @java.lang.Override public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; + @java.lang.Override public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { @@ -19670,25 +22445,33 @@ public boolean equals(final java.lang.Object obj) { } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheRequestProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheRequestProto) obj; - boolean result = true; - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; + if (!unknownFields.equals(other.unknownFields)) return false; + return true; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (19 * hash) + getDescriptor().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheRequestProto parseFrom( + java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheRequestProto parseFrom( + java.nio.ByteBuffer data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheRequestProto parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { @@ -19712,46 +22495,59 @@ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.Pu } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheRequestProto parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheRequestProto parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheRequestProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheRequestProto parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheRequestProto parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheRequestProto parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } + @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheRequestProto prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -19759,14 +22555,16 @@ protected Builder newBuilderForType( * Protobuf type {@code PurgeCacheRequestProto} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheRequestProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder implements + // @@protoc_insertion_point(builder_implements:PurgeCacheRequestProto) + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheRequestProtoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_PurgeCacheRequestProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_PurgeCacheRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -19779,36 +22577,33 @@ private Builder() { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { } } - private static Builder create() { - return new Builder(); - } - + @java.lang.Override public Builder clear() { super.clear(); return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - + @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_PurgeCacheRequestProto_descriptor; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheRequestProto getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheRequestProto.getDefaultInstance(); } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheRequestProto build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheRequestProto result = buildPartial(); if (!result.isInitialized()) { @@ -19817,12 +22612,46 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCach return result; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheRequestProto buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheRequestProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheRequestProto(this); onBuilt(); return result; } + @java.lang.Override + public Builder clone() { + return super.clone(); + } + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.setField(field, value); + } + @java.lang.Override + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + @java.lang.Override + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.addRepeatedField(field, value); + } + @java.lang.Override public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheRequestProto) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheRequestProto)other); @@ -19834,14 +22663,17 @@ public Builder mergeFrom(com.google.protobuf.Message other) { public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheRequestProto other) { if (other == org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheRequestProto.getDefaultInstance()) return this; - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } + @java.lang.Override public final boolean isInitialized() { return true; } + @java.lang.Override public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) @@ -19851,7 +22683,7 @@ public Builder mergeFrom( parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheRequestProto) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -19859,28 +22691,71 @@ public Builder mergeFrom( } return this; } + @java.lang.Override + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:PurgeCacheRequestProto) } + // @@protoc_insertion_point(class_scope:PurgeCacheRequestProto) + private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheRequestProto DEFAULT_INSTANCE; static { - defaultInstance = new PurgeCacheRequestProto(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheRequestProto(); + } + + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheRequestProto getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser + PARSER = new com.google.protobuf.AbstractParser() { + @java.lang.Override + public PurgeCacheRequestProto parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new PurgeCacheRequestProto(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheRequestProto getDefaultInstanceForType() { + return DEFAULT_INSTANCE; } - // @@protoc_insertion_point(class_scope:PurgeCacheRequestProto) } - public interface PurgeCacheResponseProtoOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface PurgeCacheResponseProtoOrBuilder extends + // @@protoc_insertion_point(interface_extends:PurgeCacheResponseProto) + com.google.protobuf.MessageOrBuilder { - // optional int64 purged_memory_bytes = 1; /** * optional int64 purged_memory_bytes = 1; + * @return Whether the purgedMemoryBytes field is set. */ boolean hasPurgedMemoryBytes(); /** * optional int64 purged_memory_bytes = 1; + * @return The purgedMemoryBytes. */ long getPurgedMemoryBytes(); } @@ -19888,35 +22763,37 @@ public interface PurgeCacheResponseProtoOrBuilder * Protobuf type {@code PurgeCacheResponseProto} */ public static final class PurgeCacheResponseProto extends - com.google.protobuf.GeneratedMessage - implements PurgeCacheResponseProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:PurgeCacheResponseProto) + PurgeCacheResponseProtoOrBuilder { + private static final long serialVersionUID = 0L; // Use PurgeCacheResponseProto.newBuilder() to construct. - private PurgeCacheResponseProto(com.google.protobuf.GeneratedMessage.Builder builder) { + private PurgeCacheResponseProto(com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private PurgeCacheResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final PurgeCacheResponseProto defaultInstance; - public static PurgeCacheResponseProto getDefaultInstance() { - return defaultInstance; + private PurgeCacheResponseProto() { } - public PurgeCacheResponseProto getDefaultInstanceForType() { - return defaultInstance; + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance( + UnusedPrivateParameter unused) { + return new PurgeCacheResponseProto(); } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private PurgeCacheResponseProto( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); @@ -19928,25 +22805,25 @@ private PurgeCacheResponseProto( case 0: done = true; break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } case 8: { bitField0_ |= 0x00000001; purgedMemoryBytes_ = input.readInt64(); break; } + default: { + if (!parseUnknownField( + input, unknownFields, extensionRegistry, tag)) { + done = true; + } + break; + } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); @@ -19957,88 +22834,69 @@ private PurgeCacheResponseProto( return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_PurgeCacheResponseProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_PurgeCacheResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheResponseProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheResponseProto.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public PurgeCacheResponseProto parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new PurgeCacheResponseProto(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - private int bitField0_; - // optional int64 purged_memory_bytes = 1; public static final int PURGED_MEMORY_BYTES_FIELD_NUMBER = 1; private long purgedMemoryBytes_; /** * optional int64 purged_memory_bytes = 1; + * @return Whether the purgedMemoryBytes field is set. */ + @java.lang.Override public boolean hasPurgedMemoryBytes() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional int64 purged_memory_bytes = 1; + * @return The purgedMemoryBytes. */ + @java.lang.Override public long getPurgedMemoryBytes() { return purgedMemoryBytes_; } - private void initFields() { - purgedMemoryBytes_ = 0L; - } private byte memoizedIsInitialized = -1; + @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } + @java.lang.Override public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); - if (((bitField0_ & 0x00000001) == 0x00000001)) { + if (((bitField0_ & 0x00000001) != 0)) { output.writeInt64(1, purgedMemoryBytes_); } - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; + @java.lang.Override public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { + if (((bitField0_ & 0x00000001) != 0)) { size += com.google.protobuf.CodedOutputStream .computeInt64Size(1, purgedMemoryBytes_); } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { @@ -20049,34 +22907,43 @@ public boolean equals(final java.lang.Object obj) { } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheResponseProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheResponseProto) obj; - boolean result = true; - result = result && (hasPurgedMemoryBytes() == other.hasPurgedMemoryBytes()); + if (hasPurgedMemoryBytes() != other.hasPurgedMemoryBytes()) return false; if (hasPurgedMemoryBytes()) { - result = result && (getPurgedMemoryBytes() - == other.getPurgedMemoryBytes()); + if (getPurgedMemoryBytes() + != other.getPurgedMemoryBytes()) return false; } - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; + if (!unknownFields.equals(other.unknownFields)) return false; + return true; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); + hash = (19 * hash) + getDescriptor().hashCode(); if (hasPurgedMemoryBytes()) { hash = (37 * hash) + PURGED_MEMORY_BYTES_FIELD_NUMBER; - hash = (53 * hash) + hashLong(getPurgedMemoryBytes()); + hash = (53 * hash) + com.google.protobuf.Internal.hashLong( + getPurgedMemoryBytes()); } - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheResponseProto parseFrom( + java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheResponseProto parseFrom( + java.nio.ByteBuffer data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheResponseProto parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { @@ -20100,46 +22967,59 @@ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.Pu } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheResponseProto parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheResponseProto parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheResponseProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheResponseProto parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheResponseProto parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheResponseProto parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } + @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheResponseProto prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -20147,14 +23027,16 @@ protected Builder newBuilderForType( * Protobuf type {@code PurgeCacheResponseProto} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheResponseProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder implements + // @@protoc_insertion_point(builder_implements:PurgeCacheResponseProto) + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheResponseProtoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_PurgeCacheResponseProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_PurgeCacheResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -20167,38 +23049,35 @@ private Builder() { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { } } - private static Builder create() { - return new Builder(); - } - + @java.lang.Override public Builder clear() { super.clear(); purgedMemoryBytes_ = 0L; bitField0_ = (bitField0_ & ~0x00000001); - return this; - } - - public Builder clone() { - return create().mergeFrom(buildPartial()); + return this; } + @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_PurgeCacheResponseProto_descriptor; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheResponseProto getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheResponseProto.getDefaultInstance(); } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheResponseProto build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheResponseProto result = buildPartial(); if (!result.isInitialized()) { @@ -20207,19 +23086,53 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCach return result; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheResponseProto buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheResponseProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheResponseProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { + if (((from_bitField0_ & 0x00000001) != 0)) { + result.purgedMemoryBytes_ = purgedMemoryBytes_; to_bitField0_ |= 0x00000001; } - result.purgedMemoryBytes_ = purgedMemoryBytes_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } + @java.lang.Override + public Builder clone() { + return super.clone(); + } + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.setField(field, value); + } + @java.lang.Override + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + @java.lang.Override + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.addRepeatedField(field, value); + } + @java.lang.Override public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheResponseProto) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheResponseProto)other); @@ -20234,14 +23147,17 @@ public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtoc if (other.hasPurgedMemoryBytes()) { setPurgedMemoryBytes(other.getPurgedMemoryBytes()); } - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } + @java.lang.Override public final boolean isInitialized() { return true; } + @java.lang.Override public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) @@ -20251,7 +23167,7 @@ public Builder mergeFrom( parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheResponseProto) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -20261,22 +23177,27 @@ public Builder mergeFrom( } private int bitField0_; - // optional int64 purged_memory_bytes = 1; private long purgedMemoryBytes_ ; /** * optional int64 purged_memory_bytes = 1; + * @return Whether the purgedMemoryBytes field is set. */ + @java.lang.Override public boolean hasPurgedMemoryBytes() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional int64 purged_memory_bytes = 1; + * @return The purgedMemoryBytes. */ + @java.lang.Override public long getPurgedMemoryBytes() { return purgedMemoryBytes_; } /** * optional int64 purged_memory_bytes = 1; + * @param value The purgedMemoryBytes to set. + * @return This builder for chaining. */ public Builder setPurgedMemoryBytes(long value) { bitField0_ |= 0x00000001; @@ -20286,6 +23207,7 @@ public Builder setPurgedMemoryBytes(long value) { } /** * optional int64 purged_memory_bytes = 1; + * @return This builder for chaining. */ public Builder clearPurgedMemoryBytes() { bitField0_ = (bitField0_ & ~0x00000001); @@ -20293,43 +23215,88 @@ public Builder clearPurgedMemoryBytes() { onChanged(); return this; } + @java.lang.Override + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:PurgeCacheResponseProto) } + // @@protoc_insertion_point(class_scope:PurgeCacheResponseProto) + private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheResponseProto DEFAULT_INSTANCE; static { - defaultInstance = new PurgeCacheResponseProto(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheResponseProto(); + } + + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheResponseProto getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser + PARSER = new com.google.protobuf.AbstractParser() { + @java.lang.Override + public PurgeCacheResponseProto parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new PurgeCacheResponseProto(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.PurgeCacheResponseProto getDefaultInstanceForType() { + return DEFAULT_INSTANCE; } - // @@protoc_insertion_point(class_scope:PurgeCacheResponseProto) } - public interface MapEntryOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface MapEntryOrBuilder extends + // @@protoc_insertion_point(interface_extends:MapEntry) + com.google.protobuf.MessageOrBuilder { - // optional string key = 1; /** * optional string key = 1; + * @return Whether the key field is set. */ boolean hasKey(); /** * optional string key = 1; + * @return The key. */ java.lang.String getKey(); /** * optional string key = 1; + * @return The bytes for key. */ com.google.protobuf.ByteString getKeyBytes(); - // optional int64 value = 2; /** * optional int64 value = 2; + * @return Whether the value field is set. */ boolean hasValue(); /** * optional int64 value = 2; + * @return The value. */ long getValue(); } @@ -20337,35 +23304,38 @@ public interface MapEntryOrBuilder * Protobuf type {@code MapEntry} */ public static final class MapEntry extends - com.google.protobuf.GeneratedMessage - implements MapEntryOrBuilder { + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:MapEntry) + MapEntryOrBuilder { + private static final long serialVersionUID = 0L; // Use MapEntry.newBuilder() to construct. - private MapEntry(com.google.protobuf.GeneratedMessage.Builder builder) { + private MapEntry(com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private MapEntry(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final MapEntry defaultInstance; - public static MapEntry getDefaultInstance() { - return defaultInstance; + private MapEntry() { + key_ = ""; } - public MapEntry getDefaultInstanceForType() { - return defaultInstance; + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance( + UnusedPrivateParameter unused) { + return new MapEntry(); } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private MapEntry( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); @@ -20377,16 +23347,10 @@ private MapEntry( case 0: done = true; break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } case 10: { + com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000001; - key_ = input.readBytes(); + key_ = bs; break; } case 16: { @@ -20394,13 +23358,20 @@ private MapEntry( value_ = input.readInt64(); break; } + default: { + if (!parseUnknownField( + input, unknownFields, extensionRegistry, tag)) { + done = true; + } + break; + } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); @@ -20411,41 +23382,30 @@ private MapEntry( return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_MapEntry_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_MapEntry_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.MapEntry.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.MapEntry.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public MapEntry parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new MapEntry(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - private int bitField0_; - // optional string key = 1; public static final int KEY_FIELD_NUMBER = 1; - private java.lang.Object key_; + private volatile java.lang.Object key_; /** * optional string key = 1; + * @return Whether the key field is set. */ + @java.lang.Override public boolean hasKey() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional string key = 1; + * @return The key. */ + @java.lang.Override public java.lang.String getKey() { java.lang.Object ref = key_; if (ref instanceof java.lang.String) { @@ -20462,7 +23422,9 @@ public java.lang.String getKey() { } /** * optional string key = 1; + * @return The bytes for key. */ + @java.lang.Override public com.google.protobuf.ByteString getKeyBytes() { java.lang.Object ref = key_; @@ -20477,73 +23439,66 @@ public java.lang.String getKey() { } } - // optional int64 value = 2; public static final int VALUE_FIELD_NUMBER = 2; private long value_; /** * optional int64 value = 2; + * @return Whether the value field is set. */ + @java.lang.Override public boolean hasValue() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * optional int64 value = 2; + * @return The value. */ + @java.lang.Override public long getValue() { return value_; } - private void initFields() { - key_ = ""; - value_ = 0L; - } private byte memoizedIsInitialized = -1; + @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } + @java.lang.Override public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); - if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeBytes(1, getKeyBytes()); + if (((bitField0_ & 0x00000001) != 0)) { + com.google.protobuf.GeneratedMessageV3.writeString(output, 1, key_); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { + if (((bitField0_ & 0x00000002) != 0)) { output.writeInt64(2, value_); } - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; + @java.lang.Override public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(1, getKeyBytes()); + if (((bitField0_ & 0x00000001) != 0)) { + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, key_); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { + if (((bitField0_ & 0x00000002) != 0)) { size += com.google.protobuf.CodedOutputStream .computeInt64Size(2, value_); } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { @@ -20554,43 +23509,52 @@ public boolean equals(final java.lang.Object obj) { } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.MapEntry other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.MapEntry) obj; - boolean result = true; - result = result && (hasKey() == other.hasKey()); + if (hasKey() != other.hasKey()) return false; if (hasKey()) { - result = result && getKey() - .equals(other.getKey()); + if (!getKey() + .equals(other.getKey())) return false; } - result = result && (hasValue() == other.hasValue()); + if (hasValue() != other.hasValue()) return false; if (hasValue()) { - result = result && (getValue() - == other.getValue()); + if (getValue() + != other.getValue()) return false; } - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; + if (!unknownFields.equals(other.unknownFields)) return false; + return true; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); + hash = (19 * hash) + getDescriptor().hashCode(); if (hasKey()) { hash = (37 * hash) + KEY_FIELD_NUMBER; hash = (53 * hash) + getKey().hashCode(); } if (hasValue()) { hash = (37 * hash) + VALUE_FIELD_NUMBER; - hash = (53 * hash) + hashLong(getValue()); + hash = (53 * hash) + com.google.protobuf.Internal.hashLong( + getValue()); } - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.MapEntry parseFrom( + java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.MapEntry parseFrom( + java.nio.ByteBuffer data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.MapEntry parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { @@ -20614,46 +23578,59 @@ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.Ma } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.MapEntry parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.MapEntry parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.MapEntry parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.MapEntry parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.MapEntry parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.MapEntry parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } + @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.MapEntry prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -20661,14 +23638,16 @@ protected Builder newBuilderForType( * Protobuf type {@code MapEntry} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.MapEntryOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder implements + // @@protoc_insertion_point(builder_implements:MapEntry) + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.MapEntryOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_MapEntry_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_MapEntry_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -20681,18 +23660,16 @@ private Builder() { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { } } - private static Builder create() { - return new Builder(); - } - + @java.lang.Override public Builder clear() { super.clear(); key_ = ""; @@ -20702,19 +23679,18 @@ public Builder clear() { return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - + @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_MapEntry_descriptor; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.MapEntry getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.MapEntry.getDefaultInstance(); } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.MapEntry build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.MapEntry result = buildPartial(); if (!result.isInitialized()) { @@ -20723,23 +23699,57 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.MapEntry return result; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.MapEntry buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.MapEntry result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.MapEntry(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { + if (((from_bitField0_ & 0x00000001) != 0)) { to_bitField0_ |= 0x00000001; } result.key_ = key_; - if (((from_bitField0_ & 0x00000002) == 0x00000002)) { + if (((from_bitField0_ & 0x00000002) != 0)) { + result.value_ = value_; to_bitField0_ |= 0x00000002; } - result.value_ = value_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } + @java.lang.Override + public Builder clone() { + return super.clone(); + } + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.setField(field, value); + } + @java.lang.Override + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + @java.lang.Override + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.addRepeatedField(field, value); + } + @java.lang.Override public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.MapEntry) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.MapEntry)other); @@ -20759,14 +23769,17 @@ public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtoc if (other.hasValue()) { setValue(other.getValue()); } - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } + @java.lang.Override public final boolean isInitialized() { return true; } + @java.lang.Override public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) @@ -20776,7 +23789,7 @@ public Builder mergeFrom( parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.MapEntry) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -20786,23 +23799,27 @@ public Builder mergeFrom( } private int bitField0_; - // optional string key = 1; private java.lang.Object key_ = ""; /** * optional string key = 1; + * @return Whether the key field is set. */ public boolean hasKey() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional string key = 1; + * @return The key. */ public java.lang.String getKey() { java.lang.Object ref = key_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - key_ = s; + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + key_ = s; + } return s; } else { return (java.lang.String) ref; @@ -20810,6 +23827,7 @@ public java.lang.String getKey() { } /** * optional string key = 1; + * @return The bytes for key. */ public com.google.protobuf.ByteString getKeyBytes() { @@ -20826,6 +23844,8 @@ public java.lang.String getKey() { } /** * optional string key = 1; + * @param value The key to set. + * @return This builder for chaining. */ public Builder setKey( java.lang.String value) { @@ -20839,6 +23859,7 @@ public Builder setKey( } /** * optional string key = 1; + * @return This builder for chaining. */ public Builder clearKey() { bitField0_ = (bitField0_ & ~0x00000001); @@ -20848,6 +23869,8 @@ public Builder clearKey() { } /** * optional string key = 1; + * @param value The bytes for key to set. + * @return This builder for chaining. */ public Builder setKeyBytes( com.google.protobuf.ByteString value) { @@ -20860,22 +23883,27 @@ public Builder setKeyBytes( return this; } - // optional int64 value = 2; private long value_ ; /** * optional int64 value = 2; + * @return Whether the value field is set. */ + @java.lang.Override public boolean hasValue() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * optional int64 value = 2; + * @return The value. */ + @java.lang.Override public long getValue() { return value_; } /** * optional int64 value = 2; + * @param value The value to set. + * @return This builder for chaining. */ public Builder setValue(long value) { bitField0_ |= 0x00000002; @@ -20885,6 +23913,7 @@ public Builder setValue(long value) { } /** * optional int64 value = 2; + * @return This builder for chaining. */ public Builder clearValue() { bitField0_ = (bitField0_ & ~0x00000002); @@ -20892,54 +23921,98 @@ public Builder clearValue() { onChanged(); return this; } + @java.lang.Override + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:MapEntry) } + // @@protoc_insertion_point(class_scope:MapEntry) + private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.MapEntry DEFAULT_INSTANCE; static { - defaultInstance = new MapEntry(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.MapEntry(); + } + + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.MapEntry getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser + PARSER = new com.google.protobuf.AbstractParser() { + @java.lang.Override + public MapEntry parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new MapEntry(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.MapEntry getDefaultInstanceForType() { + return DEFAULT_INSTANCE; } - // @@protoc_insertion_point(class_scope:MapEntry) } - public interface GetDaemonMetricsRequestProtoOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface GetDaemonMetricsRequestProtoOrBuilder extends + // @@protoc_insertion_point(interface_extends:GetDaemonMetricsRequestProto) + com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code GetDaemonMetricsRequestProto} */ public static final class GetDaemonMetricsRequestProto extends - com.google.protobuf.GeneratedMessage - implements GetDaemonMetricsRequestProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:GetDaemonMetricsRequestProto) + GetDaemonMetricsRequestProtoOrBuilder { + private static final long serialVersionUID = 0L; // Use GetDaemonMetricsRequestProto.newBuilder() to construct. - private GetDaemonMetricsRequestProto(com.google.protobuf.GeneratedMessage.Builder builder) { + private GetDaemonMetricsRequestProto(com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private GetDaemonMetricsRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final GetDaemonMetricsRequestProto defaultInstance; - public static GetDaemonMetricsRequestProto getDefaultInstance() { - return defaultInstance; + private GetDaemonMetricsRequestProto() { } - public GetDaemonMetricsRequestProto getDefaultInstanceForType() { - return defaultInstance; + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance( + UnusedPrivateParameter unused) { + return new GetDaemonMetricsRequestProto(); } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private GetDaemonMetricsRequestProto( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); try { @@ -20951,8 +24024,8 @@ private GetDaemonMetricsRequestProto( done = true; break; default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { + if (!parseUnknownField( + input, unknownFields, extensionRegistry, tag)) { done = true; } break; @@ -20963,7 +24036,7 @@ private GetDaemonMetricsRequestProto( throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); @@ -20974,63 +24047,42 @@ private GetDaemonMetricsRequestProto( return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GetDaemonMetricsRequestProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GetDaemonMetricsRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsRequestProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsRequestProto.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public GetDaemonMetricsRequestProto parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new GetDaemonMetricsRequestProto(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - - private void initFields() { - } private byte memoizedIsInitialized = -1; + @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } + @java.lang.Override public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; + @java.lang.Override public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { @@ -21041,25 +24093,33 @@ public boolean equals(final java.lang.Object obj) { } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsRequestProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsRequestProto) obj; - boolean result = true; - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; + if (!unknownFields.equals(other.unknownFields)) return false; + return true; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (19 * hash) + getDescriptor().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsRequestProto parseFrom( + java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsRequestProto parseFrom( + java.nio.ByteBuffer data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsRequestProto parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { @@ -21083,46 +24143,59 @@ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.Ge } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsRequestProto parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsRequestProto parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsRequestProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsRequestProto parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsRequestProto parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsRequestProto parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } + @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsRequestProto prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -21130,14 +24203,16 @@ protected Builder newBuilderForType( * Protobuf type {@code GetDaemonMetricsRequestProto} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsRequestProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder implements + // @@protoc_insertion_point(builder_implements:GetDaemonMetricsRequestProto) + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsRequestProtoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GetDaemonMetricsRequestProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GetDaemonMetricsRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -21150,36 +24225,33 @@ private Builder() { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { } } - private static Builder create() { - return new Builder(); - } - + @java.lang.Override public Builder clear() { super.clear(); return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - + @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GetDaemonMetricsRequestProto_descriptor; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsRequestProto getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsRequestProto.getDefaultInstance(); } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsRequestProto build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsRequestProto result = buildPartial(); if (!result.isInitialized()) { @@ -21188,12 +24260,46 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemon return result; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsRequestProto buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsRequestProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsRequestProto(this); onBuilt(); return result; } + @java.lang.Override + public Builder clone() { + return super.clone(); + } + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.setField(field, value); + } + @java.lang.Override + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + @java.lang.Override + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.addRepeatedField(field, value); + } + @java.lang.Override public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsRequestProto) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsRequestProto)other); @@ -21205,14 +24311,17 @@ public Builder mergeFrom(com.google.protobuf.Message other) { public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsRequestProto other) { if (other == org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsRequestProto.getDefaultInstance()) return this; - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } + @java.lang.Override public final boolean isInitialized() { return true; } + @java.lang.Override public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) @@ -21222,7 +24331,7 @@ public Builder mergeFrom( parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsRequestProto) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -21230,22 +24339,63 @@ public Builder mergeFrom( } return this; } + @java.lang.Override + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:GetDaemonMetricsRequestProto) } + // @@protoc_insertion_point(class_scope:GetDaemonMetricsRequestProto) + private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsRequestProto DEFAULT_INSTANCE; static { - defaultInstance = new GetDaemonMetricsRequestProto(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsRequestProto(); + } + + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsRequestProto getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser + PARSER = new com.google.protobuf.AbstractParser() { + @java.lang.Override + public GetDaemonMetricsRequestProto parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new GetDaemonMetricsRequestProto(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsRequestProto getDefaultInstanceForType() { + return DEFAULT_INSTANCE; } - // @@protoc_insertion_point(class_scope:GetDaemonMetricsRequestProto) } - public interface GetDaemonMetricsResponseProtoOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface GetDaemonMetricsResponseProtoOrBuilder extends + // @@protoc_insertion_point(interface_extends:GetDaemonMetricsResponseProto) + com.google.protobuf.MessageOrBuilder { - // repeated .MapEntry metrics = 1; /** * repeated .MapEntry metrics = 1; */ @@ -21274,35 +24424,38 @@ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.MapEntryOrBuilde * Protobuf type {@code GetDaemonMetricsResponseProto} */ public static final class GetDaemonMetricsResponseProto extends - com.google.protobuf.GeneratedMessage - implements GetDaemonMetricsResponseProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:GetDaemonMetricsResponseProto) + GetDaemonMetricsResponseProtoOrBuilder { + private static final long serialVersionUID = 0L; // Use GetDaemonMetricsResponseProto.newBuilder() to construct. - private GetDaemonMetricsResponseProto(com.google.protobuf.GeneratedMessage.Builder builder) { + private GetDaemonMetricsResponseProto(com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private GetDaemonMetricsResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final GetDaemonMetricsResponseProto defaultInstance; - public static GetDaemonMetricsResponseProto getDefaultInstance() { - return defaultInstance; + private GetDaemonMetricsResponseProto() { + metrics_ = java.util.Collections.emptyList(); } - public GetDaemonMetricsResponseProto getDefaultInstanceForType() { - return defaultInstance; + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance( + UnusedPrivateParameter unused) { + return new GetDaemonMetricsResponseProto(); } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private GetDaemonMetricsResponseProto( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); @@ -21314,19 +24467,20 @@ private GetDaemonMetricsResponseProto( case 0: done = true; break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } case 10: { - if (!((mutable_bitField0_ & 0x00000001) == 0x00000001)) { + if (!((mutable_bitField0_ & 0x00000001) != 0)) { metrics_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000001; } - metrics_.add(input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.MapEntry.PARSER, extensionRegistry)); + metrics_.add( + input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.MapEntry.PARSER, extensionRegistry)); + break; + } + default: { + if (!parseUnknownField( + input, unknownFields, extensionRegistry, tag)) { + done = true; + } break; } } @@ -21335,9 +24489,9 @@ private GetDaemonMetricsResponseProto( throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { - if (((mutable_bitField0_ & 0x00000001) == 0x00000001)) { + if (((mutable_bitField0_ & 0x00000001) != 0)) { metrics_ = java.util.Collections.unmodifiableList(metrics_); } this.unknownFields = unknownFields.build(); @@ -21349,40 +24503,27 @@ private GetDaemonMetricsResponseProto( return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GetDaemonMetricsResponseProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GetDaemonMetricsResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsResponseProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsResponseProto.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public GetDaemonMetricsResponseProto parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new GetDaemonMetricsResponseProto(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - - // repeated .MapEntry metrics = 1; public static final int METRICS_FIELD_NUMBER = 1; private java.util.List metrics_; /** * repeated .MapEntry metrics = 1; */ + @java.lang.Override public java.util.List getMetricsList() { return metrics_; } /** * repeated .MapEntry metrics = 1; */ + @java.lang.Override public java.util.List getMetricsOrBuilderList() { return metrics_; @@ -21390,47 +24531,49 @@ public java.util.Listrepeated .MapEntry metrics = 1; */ + @java.lang.Override public int getMetricsCount() { return metrics_.size(); } /** * repeated .MapEntry metrics = 1; */ + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.MapEntry getMetrics(int index) { return metrics_.get(index); } /** * repeated .MapEntry metrics = 1; */ + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.MapEntryOrBuilder getMetricsOrBuilder( int index) { return metrics_.get(index); } - private void initFields() { - metrics_ = java.util.Collections.emptyList(); - } private byte memoizedIsInitialized = -1; + @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } + @java.lang.Override public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); for (int i = 0; i < metrics_.size(); i++) { output.writeMessage(1, metrics_.get(i)); } - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; + @java.lang.Override public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; @@ -21438,18 +24581,11 @@ public int getSerializedSize() { size += com.google.protobuf.CodedOutputStream .computeMessageSize(1, metrics_.get(i)); } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { @@ -21460,31 +24596,39 @@ public boolean equals(final java.lang.Object obj) { } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsResponseProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsResponseProto) obj; - boolean result = true; - result = result && getMetricsList() - .equals(other.getMetricsList()); - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; + if (!getMetricsList() + .equals(other.getMetricsList())) return false; + if (!unknownFields.equals(other.unknownFields)) return false; + return true; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); + hash = (19 * hash) + getDescriptor().hashCode(); if (getMetricsCount() > 0) { hash = (37 * hash) + METRICS_FIELD_NUMBER; hash = (53 * hash) + getMetricsList().hashCode(); } - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsResponseProto parseFrom( + java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsResponseProto parseFrom( + java.nio.ByteBuffer data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsResponseProto parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { @@ -21508,46 +24652,59 @@ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.Ge } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsResponseProto parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsResponseProto parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsResponseProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsResponseProto parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsResponseProto parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsResponseProto parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } + @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsResponseProto prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -21555,14 +24712,16 @@ protected Builder newBuilderForType( * Protobuf type {@code GetDaemonMetricsResponseProto} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsResponseProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder implements + // @@protoc_insertion_point(builder_implements:GetDaemonMetricsResponseProto) + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsResponseProtoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GetDaemonMetricsResponseProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GetDaemonMetricsResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -21575,19 +24734,17 @@ private Builder() { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { getMetricsFieldBuilder(); } } - private static Builder create() { - return new Builder(); - } - + @java.lang.Override public Builder clear() { super.clear(); if (metricsBuilder_ == null) { @@ -21599,19 +24756,18 @@ public Builder clear() { return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - + @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GetDaemonMetricsResponseProto_descriptor; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsResponseProto getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsResponseProto.getDefaultInstance(); } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsResponseProto build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsResponseProto result = buildPartial(); if (!result.isInitialized()) { @@ -21620,11 +24776,12 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemon return result; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsResponseProto buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsResponseProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsResponseProto(this); int from_bitField0_ = bitField0_; if (metricsBuilder_ == null) { - if (((bitField0_ & 0x00000001) == 0x00000001)) { + if (((bitField0_ & 0x00000001) != 0)) { metrics_ = java.util.Collections.unmodifiableList(metrics_); bitField0_ = (bitField0_ & ~0x00000001); } @@ -21636,6 +24793,39 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemon return result; } + @java.lang.Override + public Builder clone() { + return super.clone(); + } + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.setField(field, value); + } + @java.lang.Override + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + @java.lang.Override + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.addRepeatedField(field, value); + } + @java.lang.Override public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsResponseProto) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsResponseProto)other); @@ -21666,21 +24856,24 @@ public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtoc metrics_ = other.metrics_; bitField0_ = (bitField0_ & ~0x00000001); metricsBuilder_ = - com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? + com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getMetricsFieldBuilder() : null; } else { metricsBuilder_.addAllMessages(other.metrics_); } } } - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } + @java.lang.Override public final boolean isInitialized() { return true; } + @java.lang.Override public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) @@ -21690,7 +24883,7 @@ public Builder mergeFrom( parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsResponseProto) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -21700,17 +24893,16 @@ public Builder mergeFrom( } private int bitField0_; - // repeated .MapEntry metrics = 1; private java.util.List metrics_ = java.util.Collections.emptyList(); private void ensureMetricsIsMutable() { - if (!((bitField0_ & 0x00000001) == 0x00000001)) { + if (!((bitField0_ & 0x00000001) != 0)) { metrics_ = new java.util.ArrayList(metrics_); bitField0_ |= 0x00000001; } } - private com.google.protobuf.RepeatedFieldBuilder< + private com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.MapEntry, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.MapEntry.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.MapEntryOrBuilder> metricsBuilder_; /** @@ -21842,7 +25034,8 @@ public Builder addAllMetrics( java.lang.Iterable values) { if (metricsBuilder_ == null) { ensureMetricsIsMutable(); - super.addAll(values, metrics_); + com.google.protobuf.AbstractMessageLite.Builder.addAll( + values, metrics_); onChanged(); } else { metricsBuilder_.addAllMessages(values); @@ -21925,52 +25118,96 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.MapEntry. getMetricsBuilderList() { return getMetricsFieldBuilder().getBuilderList(); } - private com.google.protobuf.RepeatedFieldBuilder< + private com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.MapEntry, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.MapEntry.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.MapEntryOrBuilder> getMetricsFieldBuilder() { if (metricsBuilder_ == null) { - metricsBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< + metricsBuilder_ = new com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.MapEntry, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.MapEntry.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.MapEntryOrBuilder>( metrics_, - ((bitField0_ & 0x00000001) == 0x00000001), + ((bitField0_ & 0x00000001) != 0), getParentForChildren(), isClean()); metrics_ = null; } return metricsBuilder_; } + @java.lang.Override + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:GetDaemonMetricsResponseProto) } + // @@protoc_insertion_point(class_scope:GetDaemonMetricsResponseProto) + private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsResponseProto DEFAULT_INSTANCE; static { - defaultInstance = new GetDaemonMetricsResponseProto(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsResponseProto(); + } + + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsResponseProto getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser + PARSER = new com.google.protobuf.AbstractParser() { + @java.lang.Override + public GetDaemonMetricsResponseProto parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new GetDaemonMetricsResponseProto(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetDaemonMetricsResponseProto getDefaultInstanceForType() { + return DEFAULT_INSTANCE; } - // @@protoc_insertion_point(class_scope:GetDaemonMetricsResponseProto) } - public interface SetCapacityRequestProtoOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface SetCapacityRequestProtoOrBuilder extends + // @@protoc_insertion_point(interface_extends:SetCapacityRequestProto) + com.google.protobuf.MessageOrBuilder { - // optional int32 executorNum = 1; /** * optional int32 executorNum = 1; + * @return Whether the executorNum field is set. */ boolean hasExecutorNum(); /** * optional int32 executorNum = 1; + * @return The executorNum. */ int getExecutorNum(); - // optional int32 queueSize = 2; /** * optional int32 queueSize = 2; + * @return Whether the queueSize field is set. */ boolean hasQueueSize(); /** * optional int32 queueSize = 2; + * @return The queueSize. */ int getQueueSize(); } @@ -21978,35 +25215,37 @@ public interface SetCapacityRequestProtoOrBuilder * Protobuf type {@code SetCapacityRequestProto} */ public static final class SetCapacityRequestProto extends - com.google.protobuf.GeneratedMessage - implements SetCapacityRequestProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:SetCapacityRequestProto) + SetCapacityRequestProtoOrBuilder { + private static final long serialVersionUID = 0L; // Use SetCapacityRequestProto.newBuilder() to construct. - private SetCapacityRequestProto(com.google.protobuf.GeneratedMessage.Builder builder) { + private SetCapacityRequestProto(com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private SetCapacityRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final SetCapacityRequestProto defaultInstance; - public static SetCapacityRequestProto getDefaultInstance() { - return defaultInstance; + private SetCapacityRequestProto() { } - public SetCapacityRequestProto getDefaultInstanceForType() { - return defaultInstance; + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance( + UnusedPrivateParameter unused) { + return new SetCapacityRequestProto(); } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private SetCapacityRequestProto( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); @@ -22018,13 +25257,6 @@ private SetCapacityRequestProto( case 0: done = true; break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } case 8: { bitField0_ |= 0x00000001; executorNum_ = input.readInt32(); @@ -22035,13 +25267,20 @@ private SetCapacityRequestProto( queueSize_ = input.readInt32(); break; } + default: { + if (!parseUnknownField( + input, unknownFields, extensionRegistry, tag)) { + done = true; + } + break; + } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); @@ -22052,112 +25291,95 @@ private SetCapacityRequestProto( return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SetCapacityRequestProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SetCapacityRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityRequestProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityRequestProto.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public SetCapacityRequestProto parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new SetCapacityRequestProto(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - private int bitField0_; - // optional int32 executorNum = 1; public static final int EXECUTORNUM_FIELD_NUMBER = 1; private int executorNum_; /** * optional int32 executorNum = 1; + * @return Whether the executorNum field is set. */ + @java.lang.Override public boolean hasExecutorNum() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional int32 executorNum = 1; + * @return The executorNum. */ + @java.lang.Override public int getExecutorNum() { return executorNum_; } - // optional int32 queueSize = 2; public static final int QUEUESIZE_FIELD_NUMBER = 2; private int queueSize_; /** * optional int32 queueSize = 2; + * @return Whether the queueSize field is set. */ + @java.lang.Override public boolean hasQueueSize() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * optional int32 queueSize = 2; + * @return The queueSize. */ + @java.lang.Override public int getQueueSize() { return queueSize_; } - private void initFields() { - executorNum_ = 0; - queueSize_ = 0; - } private byte memoizedIsInitialized = -1; + @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } + @java.lang.Override public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); - if (((bitField0_ & 0x00000001) == 0x00000001)) { + if (((bitField0_ & 0x00000001) != 0)) { output.writeInt32(1, executorNum_); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { + if (((bitField0_ & 0x00000002) != 0)) { output.writeInt32(2, queueSize_); } - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; + @java.lang.Override public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { + if (((bitField0_ & 0x00000001) != 0)) { size += com.google.protobuf.CodedOutputStream .computeInt32Size(1, executorNum_); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { + if (((bitField0_ & 0x00000002) != 0)) { size += com.google.protobuf.CodedOutputStream .computeInt32Size(2, queueSize_); } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { @@ -22168,30 +25390,27 @@ public boolean equals(final java.lang.Object obj) { } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityRequestProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityRequestProto) obj; - boolean result = true; - result = result && (hasExecutorNum() == other.hasExecutorNum()); + if (hasExecutorNum() != other.hasExecutorNum()) return false; if (hasExecutorNum()) { - result = result && (getExecutorNum() - == other.getExecutorNum()); + if (getExecutorNum() + != other.getExecutorNum()) return false; } - result = result && (hasQueueSize() == other.hasQueueSize()); + if (hasQueueSize() != other.hasQueueSize()) return false; if (hasQueueSize()) { - result = result && (getQueueSize() - == other.getQueueSize()); + if (getQueueSize() + != other.getQueueSize()) return false; } - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; + if (!unknownFields.equals(other.unknownFields)) return false; + return true; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); + hash = (19 * hash) + getDescriptor().hashCode(); if (hasExecutorNum()) { hash = (37 * hash) + EXECUTORNUM_FIELD_NUMBER; hash = (53 * hash) + getExecutorNum(); @@ -22200,11 +25419,22 @@ public int hashCode() { hash = (37 * hash) + QUEUESIZE_FIELD_NUMBER; hash = (53 * hash) + getQueueSize(); } - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityRequestProto parseFrom( + java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityRequestProto parseFrom( + java.nio.ByteBuffer data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityRequestProto parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { @@ -22228,46 +25458,59 @@ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.Se } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityRequestProto parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityRequestProto parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityRequestProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityRequestProto parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityRequestProto parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityRequestProto parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } + @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityRequestProto prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -22275,14 +25518,16 @@ protected Builder newBuilderForType( * Protobuf type {@code SetCapacityRequestProto} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityRequestProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder implements + // @@protoc_insertion_point(builder_implements:SetCapacityRequestProto) + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityRequestProtoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SetCapacityRequestProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SetCapacityRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -22295,18 +25540,16 @@ private Builder() { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { } } - private static Builder create() { - return new Builder(); - } - + @java.lang.Override public Builder clear() { super.clear(); executorNum_ = 0; @@ -22316,19 +25559,18 @@ public Builder clear() { return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - + @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SetCapacityRequestProto_descriptor; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityRequestProto getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityRequestProto.getDefaultInstance(); } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityRequestProto build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityRequestProto result = buildPartial(); if (!result.isInitialized()) { @@ -22337,23 +25579,57 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapaci return result; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityRequestProto buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityRequestProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityRequestProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { + if (((from_bitField0_ & 0x00000001) != 0)) { + result.executorNum_ = executorNum_; to_bitField0_ |= 0x00000001; } - result.executorNum_ = executorNum_; - if (((from_bitField0_ & 0x00000002) == 0x00000002)) { + if (((from_bitField0_ & 0x00000002) != 0)) { + result.queueSize_ = queueSize_; to_bitField0_ |= 0x00000002; } - result.queueSize_ = queueSize_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } + @java.lang.Override + public Builder clone() { + return super.clone(); + } + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.setField(field, value); + } + @java.lang.Override + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + @java.lang.Override + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.addRepeatedField(field, value); + } + @java.lang.Override public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityRequestProto) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityRequestProto)other); @@ -22371,14 +25647,17 @@ public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtoc if (other.hasQueueSize()) { setQueueSize(other.getQueueSize()); } - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } + @java.lang.Override public final boolean isInitialized() { return true; } + @java.lang.Override public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) @@ -22388,7 +25667,7 @@ public Builder mergeFrom( parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityRequestProto) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -22398,22 +25677,27 @@ public Builder mergeFrom( } private int bitField0_; - // optional int32 executorNum = 1; private int executorNum_ ; /** * optional int32 executorNum = 1; + * @return Whether the executorNum field is set. */ + @java.lang.Override public boolean hasExecutorNum() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional int32 executorNum = 1; + * @return The executorNum. */ + @java.lang.Override public int getExecutorNum() { return executorNum_; } /** * optional int32 executorNum = 1; + * @param value The executorNum to set. + * @return This builder for chaining. */ public Builder setExecutorNum(int value) { bitField0_ |= 0x00000001; @@ -22423,6 +25707,7 @@ public Builder setExecutorNum(int value) { } /** * optional int32 executorNum = 1; + * @return This builder for chaining. */ public Builder clearExecutorNum() { bitField0_ = (bitField0_ & ~0x00000001); @@ -22431,22 +25716,27 @@ public Builder clearExecutorNum() { return this; } - // optional int32 queueSize = 2; private int queueSize_ ; /** * optional int32 queueSize = 2; + * @return Whether the queueSize field is set. */ + @java.lang.Override public boolean hasQueueSize() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * optional int32 queueSize = 2; + * @return The queueSize. */ + @java.lang.Override public int getQueueSize() { return queueSize_; } /** * optional int32 queueSize = 2; + * @param value The queueSize to set. + * @return This builder for chaining. */ public Builder setQueueSize(int value) { bitField0_ |= 0x00000002; @@ -22456,6 +25746,7 @@ public Builder setQueueSize(int value) { } /** * optional int32 queueSize = 2; + * @return This builder for chaining. */ public Builder clearQueueSize() { bitField0_ = (bitField0_ & ~0x00000002); @@ -22463,54 +25754,98 @@ public Builder clearQueueSize() { onChanged(); return this; } + @java.lang.Override + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:SetCapacityRequestProto) } + // @@protoc_insertion_point(class_scope:SetCapacityRequestProto) + private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityRequestProto DEFAULT_INSTANCE; static { - defaultInstance = new SetCapacityRequestProto(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityRequestProto(); + } + + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityRequestProto getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser + PARSER = new com.google.protobuf.AbstractParser() { + @java.lang.Override + public SetCapacityRequestProto parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new SetCapacityRequestProto(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityRequestProto getDefaultInstanceForType() { + return DEFAULT_INSTANCE; } - // @@protoc_insertion_point(class_scope:SetCapacityRequestProto) } - public interface SetCapacityResponseProtoOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface SetCapacityResponseProtoOrBuilder extends + // @@protoc_insertion_point(interface_extends:SetCapacityResponseProto) + com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code SetCapacityResponseProto} */ public static final class SetCapacityResponseProto extends - com.google.protobuf.GeneratedMessage - implements SetCapacityResponseProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:SetCapacityResponseProto) + SetCapacityResponseProtoOrBuilder { + private static final long serialVersionUID = 0L; // Use SetCapacityResponseProto.newBuilder() to construct. - private SetCapacityResponseProto(com.google.protobuf.GeneratedMessage.Builder builder) { + private SetCapacityResponseProto(com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private SetCapacityResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final SetCapacityResponseProto defaultInstance; - public static SetCapacityResponseProto getDefaultInstance() { - return defaultInstance; + private SetCapacityResponseProto() { } - public SetCapacityResponseProto getDefaultInstanceForType() { - return defaultInstance; + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance( + UnusedPrivateParameter unused) { + return new SetCapacityResponseProto(); } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private SetCapacityResponseProto( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); try { @@ -22522,8 +25857,8 @@ private SetCapacityResponseProto( done = true; break; default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { + if (!parseUnknownField( + input, unknownFields, extensionRegistry, tag)) { done = true; } break; @@ -22534,7 +25869,7 @@ private SetCapacityResponseProto( throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); @@ -22545,63 +25880,42 @@ private SetCapacityResponseProto( return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SetCapacityResponseProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SetCapacityResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityResponseProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityResponseProto.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public SetCapacityResponseProto parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new SetCapacityResponseProto(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - - private void initFields() { - } private byte memoizedIsInitialized = -1; + @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } + @java.lang.Override public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; + @java.lang.Override public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { @@ -22612,25 +25926,33 @@ public boolean equals(final java.lang.Object obj) { } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityResponseProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityResponseProto) obj; - boolean result = true; - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; + if (!unknownFields.equals(other.unknownFields)) return false; + return true; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (19 * hash) + getDescriptor().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityResponseProto parseFrom( + java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityResponseProto parseFrom( + java.nio.ByteBuffer data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityResponseProto parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { @@ -22654,46 +25976,59 @@ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.Se } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityResponseProto parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityResponseProto parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityResponseProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityResponseProto parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityResponseProto parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityResponseProto parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } + @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityResponseProto prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -22701,14 +26036,16 @@ protected Builder newBuilderForType( * Protobuf type {@code SetCapacityResponseProto} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityResponseProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder implements + // @@protoc_insertion_point(builder_implements:SetCapacityResponseProto) + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityResponseProtoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SetCapacityResponseProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SetCapacityResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -22721,36 +26058,33 @@ private Builder() { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { } } - private static Builder create() { - return new Builder(); - } - + @java.lang.Override public Builder clear() { super.clear(); return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - + @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_SetCapacityResponseProto_descriptor; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityResponseProto getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityResponseProto.getDefaultInstance(); } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityResponseProto build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityResponseProto result = buildPartial(); if (!result.isInitialized()) { @@ -22759,12 +26093,46 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapaci return result; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityResponseProto buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityResponseProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityResponseProto(this); onBuilt(); return result; } + @java.lang.Override + public Builder clone() { + return super.clone(); + } + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.setField(field, value); + } + @java.lang.Override + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + @java.lang.Override + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.addRepeatedField(field, value); + } + @java.lang.Override public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityResponseProto) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityResponseProto)other); @@ -22776,14 +26144,17 @@ public Builder mergeFrom(com.google.protobuf.Message other) { public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityResponseProto other) { if (other == org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityResponseProto.getDefaultInstance()) return this; - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } + @java.lang.Override public final boolean isInitialized() { return true; } + @java.lang.Override public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) @@ -22793,7 +26164,7 @@ public Builder mergeFrom( parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityResponseProto) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -22801,37 +26172,80 @@ public Builder mergeFrom( } return this; } + @java.lang.Override + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:SetCapacityResponseProto) } + // @@protoc_insertion_point(class_scope:SetCapacityResponseProto) + private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityResponseProto DEFAULT_INSTANCE; static { - defaultInstance = new SetCapacityResponseProto(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityResponseProto(); + } + + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityResponseProto getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser + PARSER = new com.google.protobuf.AbstractParser() { + @java.lang.Override + public SetCapacityResponseProto parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new SetCapacityResponseProto(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.SetCapacityResponseProto getDefaultInstanceForType() { + return DEFAULT_INSTANCE; } - // @@protoc_insertion_point(class_scope:SetCapacityResponseProto) } - public interface EvictEntityRequestProtoOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface EvictEntityRequestProtoOrBuilder extends + // @@protoc_insertion_point(interface_extends:EvictEntityRequestProto) + com.google.protobuf.MessageOrBuilder { - // required string db_name = 1; /** * required string db_name = 1; + * @return Whether the dbName field is set. */ boolean hasDbName(); /** * required string db_name = 1; + * @return The dbName. */ java.lang.String getDbName(); /** * required string db_name = 1; + * @return The bytes for dbName. */ com.google.protobuf.ByteString getDbNameBytes(); - // repeated .TableProto table = 2; /** * repeated .TableProto table = 2; */ @@ -22857,42 +26271,46 @@ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TableProtoOrBuil int index); } /** - * Protobuf type {@code EvictEntityRequestProto} - * *
    * Used for proactive eviction request. Must contain one DB name, and optionally table information.
    * 
+ * + * Protobuf type {@code EvictEntityRequestProto} */ public static final class EvictEntityRequestProto extends - com.google.protobuf.GeneratedMessage - implements EvictEntityRequestProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:EvictEntityRequestProto) + EvictEntityRequestProtoOrBuilder { + private static final long serialVersionUID = 0L; // Use EvictEntityRequestProto.newBuilder() to construct. - private EvictEntityRequestProto(com.google.protobuf.GeneratedMessage.Builder builder) { + private EvictEntityRequestProto(com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private EvictEntityRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final EvictEntityRequestProto defaultInstance; - public static EvictEntityRequestProto getDefaultInstance() { - return defaultInstance; + private EvictEntityRequestProto() { + dbName_ = ""; + table_ = java.util.Collections.emptyList(); } - public EvictEntityRequestProto getDefaultInstanceForType() { - return defaultInstance; + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance( + UnusedPrivateParameter unused) { + return new EvictEntityRequestProto(); } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private EvictEntityRequestProto( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); @@ -22904,24 +26322,26 @@ private EvictEntityRequestProto( case 0: done = true; break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } case 10: { + com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000001; - dbName_ = input.readBytes(); + dbName_ = bs; break; } case 18: { - if (!((mutable_bitField0_ & 0x00000002) == 0x00000002)) { + if (!((mutable_bitField0_ & 0x00000002) != 0)) { table_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000002; } - table_.add(input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TableProto.PARSER, extensionRegistry)); + table_.add( + input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TableProto.PARSER, extensionRegistry)); + break; + } + default: { + if (!parseUnknownField( + input, unknownFields, extensionRegistry, tag)) { + done = true; + } break; } } @@ -22930,9 +26350,9 @@ private EvictEntityRequestProto( throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { - if (((mutable_bitField0_ & 0x00000002) == 0x00000002)) { + if (((mutable_bitField0_ & 0x00000002) != 0)) { table_ = java.util.Collections.unmodifiableList(table_); } this.unknownFields = unknownFields.build(); @@ -22944,41 +26364,30 @@ private EvictEntityRequestProto( return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_EvictEntityRequestProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_EvictEntityRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityRequestProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityRequestProto.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public EvictEntityRequestProto parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new EvictEntityRequestProto(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - private int bitField0_; - // required string db_name = 1; public static final int DB_NAME_FIELD_NUMBER = 1; - private java.lang.Object dbName_; + private volatile java.lang.Object dbName_; /** * required string db_name = 1; + * @return Whether the dbName field is set. */ + @java.lang.Override public boolean hasDbName() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * required string db_name = 1; + * @return The dbName. */ + @java.lang.Override public java.lang.String getDbName() { java.lang.Object ref = dbName_; if (ref instanceof java.lang.String) { @@ -22995,7 +26404,9 @@ public java.lang.String getDbName() { } /** * required string db_name = 1; + * @return The bytes for dbName. */ + @java.lang.Override public com.google.protobuf.ByteString getDbNameBytes() { java.lang.Object ref = dbName_; @@ -23010,18 +26421,19 @@ public java.lang.String getDbName() { } } - // repeated .TableProto table = 2; public static final int TABLE_FIELD_NUMBER = 2; private java.util.List table_; /** * repeated .TableProto table = 2; */ + @java.lang.Override public java.util.List getTableList() { return table_; } /** * repeated .TableProto table = 2; */ + @java.lang.Override public java.util.List getTableOrBuilderList() { return table_; @@ -23029,31 +26441,32 @@ public java.util.Listrepeated .TableProto table = 2; */ + @java.lang.Override public int getTableCount() { return table_.size(); } /** * repeated .TableProto table = 2; */ + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TableProto getTable(int index) { return table_.get(index); } /** * repeated .TableProto table = 2; */ + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TableProtoOrBuilder getTableOrBuilder( int index) { return table_.get(index); } - private void initFields() { - dbName_ = ""; - table_ = java.util.Collections.emptyList(); - } private byte memoizedIsInitialized = -1; + @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; if (!hasDbName()) { memoizedIsInitialized = 0; @@ -23069,44 +26482,36 @@ public final boolean isInitialized() { return true; } + @java.lang.Override public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); - if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeBytes(1, getDbNameBytes()); + if (((bitField0_ & 0x00000001) != 0)) { + com.google.protobuf.GeneratedMessageV3.writeString(output, 1, dbName_); } for (int i = 0; i < table_.size(); i++) { output.writeMessage(2, table_.get(i)); } - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; + @java.lang.Override public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(1, getDbNameBytes()); + if (((bitField0_ & 0x00000001) != 0)) { + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, dbName_); } for (int i = 0; i < table_.size(); i++) { size += com.google.protobuf.CodedOutputStream .computeMessageSize(2, table_.get(i)); } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { @@ -23117,27 +26522,24 @@ public boolean equals(final java.lang.Object obj) { } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityRequestProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityRequestProto) obj; - boolean result = true; - result = result && (hasDbName() == other.hasDbName()); + if (hasDbName() != other.hasDbName()) return false; if (hasDbName()) { - result = result && getDbName() - .equals(other.getDbName()); + if (!getDbName() + .equals(other.getDbName())) return false; } - result = result && getTableList() - .equals(other.getTableList()); - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; + if (!getTableList() + .equals(other.getTableList())) return false; + if (!unknownFields.equals(other.unknownFields)) return false; + return true; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); + hash = (19 * hash) + getDescriptor().hashCode(); if (hasDbName()) { hash = (37 * hash) + DB_NAME_FIELD_NUMBER; hash = (53 * hash) + getDbName().hashCode(); @@ -23146,11 +26548,22 @@ public int hashCode() { hash = (37 * hash) + TABLE_FIELD_NUMBER; hash = (53 * hash) + getTableList().hashCode(); } - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityRequestProto parseFrom( + java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityRequestProto parseFrom( + java.nio.ByteBuffer data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityRequestProto parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { @@ -23174,65 +26587,80 @@ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.Ev } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityRequestProto parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityRequestProto parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityRequestProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityRequestProto parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityRequestProto parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityRequestProto parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } + @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityRequestProto prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** - * Protobuf type {@code EvictEntityRequestProto} - * *
      * Used for proactive eviction request. Must contain one DB name, and optionally table information.
      * 
+ * + * Protobuf type {@code EvictEntityRequestProto} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityRequestProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder implements + // @@protoc_insertion_point(builder_implements:EvictEntityRequestProto) + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityRequestProtoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_EvictEntityRequestProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_EvictEntityRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -23245,19 +26673,17 @@ private Builder() { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { getTableFieldBuilder(); } } - private static Builder create() { - return new Builder(); - } - + @java.lang.Override public Builder clear() { super.clear(); dbName_ = ""; @@ -23271,19 +26697,18 @@ public Builder clear() { return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - + @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_EvictEntityRequestProto_descriptor; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityRequestProto getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityRequestProto.getDefaultInstance(); } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityRequestProto build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityRequestProto result = buildPartial(); if (!result.isInitialized()) { @@ -23292,16 +26717,17 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEnti return result; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityRequestProto buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityRequestProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityRequestProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { + if (((from_bitField0_ & 0x00000001) != 0)) { to_bitField0_ |= 0x00000001; } result.dbName_ = dbName_; if (tableBuilder_ == null) { - if (((bitField0_ & 0x00000002) == 0x00000002)) { + if (((bitField0_ & 0x00000002) != 0)) { table_ = java.util.Collections.unmodifiableList(table_); bitField0_ = (bitField0_ & ~0x00000002); } @@ -23314,6 +26740,39 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEnti return result; } + @java.lang.Override + public Builder clone() { + return super.clone(); + } + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.setField(field, value); + } + @java.lang.Override + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + @java.lang.Override + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.addRepeatedField(field, value); + } + @java.lang.Override public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityRequestProto) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityRequestProto)other); @@ -23349,31 +26808,32 @@ public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtoc table_ = other.table_; bitField0_ = (bitField0_ & ~0x00000002); tableBuilder_ = - com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? + com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getTableFieldBuilder() : null; } else { tableBuilder_.addAllMessages(other.table_); } } } - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } + @java.lang.Override public final boolean isInitialized() { if (!hasDbName()) { - return false; } for (int i = 0; i < getTableCount(); i++) { if (!getTable(i).isInitialized()) { - return false; } } return true; } + @java.lang.Override public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) @@ -23383,7 +26843,7 @@ public Builder mergeFrom( parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityRequestProto) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -23393,23 +26853,27 @@ public Builder mergeFrom( } private int bitField0_; - // required string db_name = 1; private java.lang.Object dbName_ = ""; /** * required string db_name = 1; + * @return Whether the dbName field is set. */ public boolean hasDbName() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * required string db_name = 1; + * @return The dbName. */ public java.lang.String getDbName() { java.lang.Object ref = dbName_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - dbName_ = s; + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + dbName_ = s; + } return s; } else { return (java.lang.String) ref; @@ -23417,6 +26881,7 @@ public java.lang.String getDbName() { } /** * required string db_name = 1; + * @return The bytes for dbName. */ public com.google.protobuf.ByteString getDbNameBytes() { @@ -23433,6 +26898,8 @@ public java.lang.String getDbName() { } /** * required string db_name = 1; + * @param value The dbName to set. + * @return This builder for chaining. */ public Builder setDbName( java.lang.String value) { @@ -23446,6 +26913,7 @@ public Builder setDbName( } /** * required string db_name = 1; + * @return This builder for chaining. */ public Builder clearDbName() { bitField0_ = (bitField0_ & ~0x00000001); @@ -23455,6 +26923,8 @@ public Builder clearDbName() { } /** * required string db_name = 1; + * @param value The bytes for dbName to set. + * @return This builder for chaining. */ public Builder setDbNameBytes( com.google.protobuf.ByteString value) { @@ -23467,17 +26937,16 @@ public Builder setDbNameBytes( return this; } - // repeated .TableProto table = 2; private java.util.List table_ = java.util.Collections.emptyList(); private void ensureTableIsMutable() { - if (!((bitField0_ & 0x00000002) == 0x00000002)) { + if (!((bitField0_ & 0x00000002) != 0)) { table_ = new java.util.ArrayList(table_); bitField0_ |= 0x00000002; } } - private com.google.protobuf.RepeatedFieldBuilder< + private com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TableProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TableProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TableProtoOrBuilder> tableBuilder_; /** @@ -23609,7 +27078,8 @@ public Builder addAllTable( java.lang.Iterable values) { if (tableBuilder_ == null) { ensureTableIsMutable(); - super.addAll(values, table_); + com.google.protobuf.AbstractMessageLite.Builder.addAll( + values, table_); onChanged(); } else { tableBuilder_.addAllMessages(values); @@ -23692,130 +27162,189 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TableProt getTableBuilderList() { return getTableFieldBuilder().getBuilderList(); } - private com.google.protobuf.RepeatedFieldBuilder< + private com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TableProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TableProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TableProtoOrBuilder> getTableFieldBuilder() { if (tableBuilder_ == null) { - tableBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< + tableBuilder_ = new com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TableProto, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TableProto.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TableProtoOrBuilder>( table_, - ((bitField0_ & 0x00000002) == 0x00000002), + ((bitField0_ & 0x00000002) != 0), getParentForChildren(), isClean()); table_ = null; } return tableBuilder_; } + @java.lang.Override + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:EvictEntityRequestProto) } + // @@protoc_insertion_point(class_scope:EvictEntityRequestProto) + private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityRequestProto DEFAULT_INSTANCE; static { - defaultInstance = new EvictEntityRequestProto(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityRequestProto(); + } + + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityRequestProto getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser + PARSER = new com.google.protobuf.AbstractParser() { + @java.lang.Override + public EvictEntityRequestProto parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new EvictEntityRequestProto(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityRequestProto getDefaultInstanceForType() { + return DEFAULT_INSTANCE; } - // @@protoc_insertion_point(class_scope:EvictEntityRequestProto) } - public interface TableProtoOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface TableProtoOrBuilder extends + // @@protoc_insertion_point(interface_extends:TableProto) + com.google.protobuf.MessageOrBuilder { - // required string table_name = 1; /** * required string table_name = 1; + * @return Whether the tableName field is set. */ boolean hasTableName(); /** * required string table_name = 1; + * @return The tableName. */ java.lang.String getTableName(); /** * required string table_name = 1; + * @return The bytes for tableName. */ com.google.protobuf.ByteString getTableNameBytes(); - // repeated string part_key = 2; /** * repeated string part_key = 2; + * @return A list containing the partKey. */ java.util.List - getPartKeyList(); + getPartKeyList(); /** * repeated string part_key = 2; + * @return The count of partKey. */ int getPartKeyCount(); /** * repeated string part_key = 2; + * @param index The index of the element to return. + * @return The partKey at the given index. */ java.lang.String getPartKey(int index); /** * repeated string part_key = 2; + * @param index The index of the value to return. + * @return The bytes of the partKey at the given index. */ com.google.protobuf.ByteString getPartKeyBytes(int index); - // repeated string part_val = 3; /** * repeated string part_val = 3; + * @return A list containing the partVal. */ java.util.List - getPartValList(); + getPartValList(); /** * repeated string part_val = 3; + * @return The count of partVal. */ int getPartValCount(); /** * repeated string part_val = 3; + * @param index The index of the element to return. + * @return The partVal at the given index. */ java.lang.String getPartVal(int index); /** * repeated string part_val = 3; + * @param index The index of the value to return. + * @return The bytes of the partVal at the given index. */ com.google.protobuf.ByteString getPartValBytes(int index); } /** - * Protobuf type {@code TableProto} - * *
    * Used in EvictEntityRequestProto, can be used for non-partitioned and partitioned tables too.
    * For the latter part_key contains only the keys, part_val has the values for all partitions on all keys:
    * e.g.: for partitions pk0=p00/pk1=p01/pk2=p02 and pk0=p10/pk1=p11/pk2=p12
    * part_key: [pk0, pk1, pk2], part_val: [p00, p01, p02, p10, p11, p12]
    * 
+ * + * Protobuf type {@code TableProto} */ public static final class TableProto extends - com.google.protobuf.GeneratedMessage - implements TableProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:TableProto) + TableProtoOrBuilder { + private static final long serialVersionUID = 0L; // Use TableProto.newBuilder() to construct. - private TableProto(com.google.protobuf.GeneratedMessage.Builder builder) { + private TableProto(com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private TableProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final TableProto defaultInstance; - public static TableProto getDefaultInstance() { - return defaultInstance; + private TableProto() { + tableName_ = ""; + partKey_ = com.google.protobuf.LazyStringArrayList.EMPTY; + partVal_ = com.google.protobuf.LazyStringArrayList.EMPTY; } - public TableProto getDefaultInstanceForType() { - return defaultInstance; + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance( + UnusedPrivateParameter unused) { + return new TableProto(); } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private TableProto( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); @@ -23827,32 +27356,35 @@ private TableProto( case 0: done = true; break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } case 10: { + com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000001; - tableName_ = input.readBytes(); + tableName_ = bs; break; } case 18: { - if (!((mutable_bitField0_ & 0x00000002) == 0x00000002)) { + com.google.protobuf.ByteString bs = input.readBytes(); + if (!((mutable_bitField0_ & 0x00000002) != 0)) { partKey_ = new com.google.protobuf.LazyStringArrayList(); mutable_bitField0_ |= 0x00000002; } - partKey_.add(input.readBytes()); + partKey_.add(bs); break; } case 26: { - if (!((mutable_bitField0_ & 0x00000004) == 0x00000004)) { + com.google.protobuf.ByteString bs = input.readBytes(); + if (!((mutable_bitField0_ & 0x00000004) != 0)) { partVal_ = new com.google.protobuf.LazyStringArrayList(); mutable_bitField0_ |= 0x00000004; } - partVal_.add(input.readBytes()); + partVal_.add(bs); + break; + } + default: { + if (!parseUnknownField( + input, unknownFields, extensionRegistry, tag)) { + done = true; + } break; } } @@ -23861,13 +27393,13 @@ private TableProto( throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { - if (((mutable_bitField0_ & 0x00000002) == 0x00000002)) { - partKey_ = new com.google.protobuf.UnmodifiableLazyStringList(partKey_); + if (((mutable_bitField0_ & 0x00000002) != 0)) { + partKey_ = partKey_.getUnmodifiableView(); } - if (((mutable_bitField0_ & 0x00000004) == 0x00000004)) { - partVal_ = new com.google.protobuf.UnmodifiableLazyStringList(partVal_); + if (((mutable_bitField0_ & 0x00000004) != 0)) { + partVal_ = partVal_.getUnmodifiableView(); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); @@ -23878,41 +27410,30 @@ private TableProto( return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_TableProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_TableProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TableProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TableProto.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public TableProto parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new TableProto(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - private int bitField0_; - // required string table_name = 1; public static final int TABLE_NAME_FIELD_NUMBER = 1; - private java.lang.Object tableName_; + private volatile java.lang.Object tableName_; /** * required string table_name = 1; + * @return Whether the tableName field is set. */ + @java.lang.Override public boolean hasTableName() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * required string table_name = 1; + * @return The tableName. */ + @java.lang.Override public java.lang.String getTableName() { java.lang.Object ref = tableName_; if (ref instanceof java.lang.String) { @@ -23929,7 +27450,9 @@ public java.lang.String getTableName() { } /** * required string table_name = 1; + * @return The bytes for tableName. */ + @java.lang.Override public com.google.protobuf.ByteString getTableNameBytes() { java.lang.Object ref = tableName_; @@ -23944,75 +27467,82 @@ public java.lang.String getTableName() { } } - // repeated string part_key = 2; public static final int PART_KEY_FIELD_NUMBER = 2; private com.google.protobuf.LazyStringList partKey_; /** * repeated string part_key = 2; + * @return A list containing the partKey. */ - public java.util.List + public com.google.protobuf.ProtocolStringList getPartKeyList() { return partKey_; } /** * repeated string part_key = 2; + * @return The count of partKey. */ public int getPartKeyCount() { return partKey_.size(); } /** * repeated string part_key = 2; + * @param index The index of the element to return. + * @return The partKey at the given index. */ public java.lang.String getPartKey(int index) { return partKey_.get(index); } /** * repeated string part_key = 2; + * @param index The index of the value to return. + * @return The bytes of the partKey at the given index. */ public com.google.protobuf.ByteString getPartKeyBytes(int index) { return partKey_.getByteString(index); } - // repeated string part_val = 3; public static final int PART_VAL_FIELD_NUMBER = 3; private com.google.protobuf.LazyStringList partVal_; /** * repeated string part_val = 3; + * @return A list containing the partVal. */ - public java.util.List + public com.google.protobuf.ProtocolStringList getPartValList() { return partVal_; } /** * repeated string part_val = 3; + * @return The count of partVal. */ public int getPartValCount() { return partVal_.size(); } /** * repeated string part_val = 3; + * @param index The index of the element to return. + * @return The partVal at the given index. */ public java.lang.String getPartVal(int index) { return partVal_.get(index); } /** * repeated string part_val = 3; + * @param index The index of the value to return. + * @return The bytes of the partVal at the given index. */ public com.google.protobuf.ByteString getPartValBytes(int index) { return partVal_.getByteString(index); } - private void initFields() { - tableName_ = ""; - partKey_ = com.google.protobuf.LazyStringArrayList.EMPTY; - partVal_ = com.google.protobuf.LazyStringArrayList.EMPTY; - } private byte memoizedIsInitialized = -1; + @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; if (!hasTableName()) { memoizedIsInitialized = 0; @@ -24022,36 +27552,34 @@ public final boolean isInitialized() { return true; } + @java.lang.Override public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); - if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeBytes(1, getTableNameBytes()); + if (((bitField0_ & 0x00000001) != 0)) { + com.google.protobuf.GeneratedMessageV3.writeString(output, 1, tableName_); } for (int i = 0; i < partKey_.size(); i++) { - output.writeBytes(2, partKey_.getByteString(i)); + com.google.protobuf.GeneratedMessageV3.writeString(output, 2, partKey_.getRaw(i)); } for (int i = 0; i < partVal_.size(); i++) { - output.writeBytes(3, partVal_.getByteString(i)); + com.google.protobuf.GeneratedMessageV3.writeString(output, 3, partVal_.getRaw(i)); } - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; + @java.lang.Override public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(1, getTableNameBytes()); + if (((bitField0_ & 0x00000001) != 0)) { + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, tableName_); } { int dataSize = 0; for (int i = 0; i < partKey_.size(); i++) { - dataSize += com.google.protobuf.CodedOutputStream - .computeBytesSizeNoTag(partKey_.getByteString(i)); + dataSize += computeStringSizeNoTag(partKey_.getRaw(i)); } size += dataSize; size += 1 * getPartKeyList().size(); @@ -24059,24 +27587,16 @@ public int getSerializedSize() { { int dataSize = 0; for (int i = 0; i < partVal_.size(); i++) { - dataSize += com.google.protobuf.CodedOutputStream - .computeBytesSizeNoTag(partVal_.getByteString(i)); + dataSize += computeStringSizeNoTag(partVal_.getRaw(i)); } size += dataSize; size += 1 * getPartValList().size(); } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { @@ -24087,29 +27607,26 @@ public boolean equals(final java.lang.Object obj) { } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TableProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TableProto) obj; - boolean result = true; - result = result && (hasTableName() == other.hasTableName()); + if (hasTableName() != other.hasTableName()) return false; if (hasTableName()) { - result = result && getTableName() - .equals(other.getTableName()); - } - result = result && getPartKeyList() - .equals(other.getPartKeyList()); - result = result && getPartValList() - .equals(other.getPartValList()); - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; + if (!getTableName() + .equals(other.getTableName())) return false; + } + if (!getPartKeyList() + .equals(other.getPartKeyList())) return false; + if (!getPartValList() + .equals(other.getPartValList())) return false; + if (!unknownFields.equals(other.unknownFields)) return false; + return true; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); + hash = (19 * hash) + getDescriptor().hashCode(); if (hasTableName()) { hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER; hash = (53 * hash) + getTableName().hashCode(); @@ -24122,11 +27639,22 @@ public int hashCode() { hash = (37 * hash) + PART_VAL_FIELD_NUMBER; hash = (53 * hash) + getPartValList().hashCode(); } - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TableProto parseFrom( + java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TableProto parseFrom( + java.nio.ByteBuffer data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TableProto parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { @@ -24150,68 +27678,83 @@ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.Ta } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TableProto parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TableProto parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TableProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TableProto parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TableProto parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TableProto parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } + @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TableProto prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** - * Protobuf type {@code TableProto} - * *
      * Used in EvictEntityRequestProto, can be used for non-partitioned and partitioned tables too.
      * For the latter part_key contains only the keys, part_val has the values for all partitions on all keys:
      * e.g.: for partitions pk0=p00/pk1=p01/pk2=p02 and pk0=p10/pk1=p11/pk2=p12
      * part_key: [pk0, pk1, pk2], part_val: [p00, p01, p02, p10, p11, p12]
      * 
+ * + * Protobuf type {@code TableProto} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TableProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder implements + // @@protoc_insertion_point(builder_implements:TableProto) + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TableProtoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_TableProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_TableProto_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -24224,18 +27767,16 @@ private Builder() { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { } } - private static Builder create() { - return new Builder(); - } - + @java.lang.Override public Builder clear() { super.clear(); tableName_ = ""; @@ -24247,19 +27788,18 @@ public Builder clear() { return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - + @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_TableProto_descriptor; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TableProto getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TableProto.getDefaultInstance(); } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TableProto build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TableProto result = buildPartial(); if (!result.isInitialized()) { @@ -24268,23 +27808,22 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TableProt return result; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TableProto buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TableProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TableProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { + if (((from_bitField0_ & 0x00000001) != 0)) { to_bitField0_ |= 0x00000001; } result.tableName_ = tableName_; - if (((bitField0_ & 0x00000002) == 0x00000002)) { - partKey_ = new com.google.protobuf.UnmodifiableLazyStringList( - partKey_); + if (((bitField0_ & 0x00000002) != 0)) { + partKey_ = partKey_.getUnmodifiableView(); bitField0_ = (bitField0_ & ~0x00000002); } result.partKey_ = partKey_; - if (((bitField0_ & 0x00000004) == 0x00000004)) { - partVal_ = new com.google.protobuf.UnmodifiableLazyStringList( - partVal_); + if (((bitField0_ & 0x00000004) != 0)) { + partVal_ = partVal_.getUnmodifiableView(); bitField0_ = (bitField0_ & ~0x00000004); } result.partVal_ = partVal_; @@ -24293,6 +27832,39 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TableProt return result; } + @java.lang.Override + public Builder clone() { + return super.clone(); + } + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.setField(field, value); + } + @java.lang.Override + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + @java.lang.Override + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.addRepeatedField(field, value); + } + @java.lang.Override public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TableProto) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TableProto)other); @@ -24329,18 +27901,20 @@ public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtoc } onChanged(); } - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } + @java.lang.Override public final boolean isInitialized() { if (!hasTableName()) { - return false; } return true; } + @java.lang.Override public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) @@ -24350,7 +27924,7 @@ public Builder mergeFrom( parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TableProto) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -24360,23 +27934,27 @@ public Builder mergeFrom( } private int bitField0_; - // required string table_name = 1; private java.lang.Object tableName_ = ""; /** * required string table_name = 1; + * @return Whether the tableName field is set. */ public boolean hasTableName() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * required string table_name = 1; + * @return The tableName. */ public java.lang.String getTableName() { java.lang.Object ref = tableName_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - tableName_ = s; + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + tableName_ = s; + } return s; } else { return (java.lang.String) ref; @@ -24384,6 +27962,7 @@ public java.lang.String getTableName() { } /** * required string table_name = 1; + * @return The bytes for tableName. */ public com.google.protobuf.ByteString getTableNameBytes() { @@ -24400,6 +27979,8 @@ public java.lang.String getTableName() { } /** * required string table_name = 1; + * @param value The tableName to set. + * @return This builder for chaining. */ public Builder setTableName( java.lang.String value) { @@ -24413,6 +27994,7 @@ public Builder setTableName( } /** * required string table_name = 1; + * @return This builder for chaining. */ public Builder clearTableName() { bitField0_ = (bitField0_ & ~0x00000001); @@ -24422,6 +28004,8 @@ public Builder clearTableName() { } /** * required string table_name = 1; + * @param value The bytes for tableName to set. + * @return This builder for chaining. */ public Builder setTableNameBytes( com.google.protobuf.ByteString value) { @@ -24434,35 +28018,40 @@ public Builder setTableNameBytes( return this; } - // repeated string part_key = 2; private com.google.protobuf.LazyStringList partKey_ = com.google.protobuf.LazyStringArrayList.EMPTY; private void ensurePartKeyIsMutable() { - if (!((bitField0_ & 0x00000002) == 0x00000002)) { + if (!((bitField0_ & 0x00000002) != 0)) { partKey_ = new com.google.protobuf.LazyStringArrayList(partKey_); bitField0_ |= 0x00000002; } } /** * repeated string part_key = 2; + * @return A list containing the partKey. */ - public java.util.List + public com.google.protobuf.ProtocolStringList getPartKeyList() { - return java.util.Collections.unmodifiableList(partKey_); + return partKey_.getUnmodifiableView(); } /** * repeated string part_key = 2; + * @return The count of partKey. */ public int getPartKeyCount() { return partKey_.size(); } /** * repeated string part_key = 2; + * @param index The index of the element to return. + * @return The partKey at the given index. */ public java.lang.String getPartKey(int index) { return partKey_.get(index); } /** * repeated string part_key = 2; + * @param index The index of the value to return. + * @return The bytes of the partKey at the given index. */ public com.google.protobuf.ByteString getPartKeyBytes(int index) { @@ -24470,6 +28059,9 @@ public java.lang.String getPartKey(int index) { } /** * repeated string part_key = 2; + * @param index The index to set the value at. + * @param value The partKey to set. + * @return This builder for chaining. */ public Builder setPartKey( int index, java.lang.String value) { @@ -24483,6 +28075,8 @@ public Builder setPartKey( } /** * repeated string part_key = 2; + * @param value The partKey to add. + * @return This builder for chaining. */ public Builder addPartKey( java.lang.String value) { @@ -24496,16 +28090,20 @@ public Builder addPartKey( } /** * repeated string part_key = 2; + * @param values The partKey to add. + * @return This builder for chaining. */ public Builder addAllPartKey( java.lang.Iterable values) { ensurePartKeyIsMutable(); - super.addAll(values, partKey_); + com.google.protobuf.AbstractMessageLite.Builder.addAll( + values, partKey_); onChanged(); return this; } /** * repeated string part_key = 2; + * @return This builder for chaining. */ public Builder clearPartKey() { partKey_ = com.google.protobuf.LazyStringArrayList.EMPTY; @@ -24515,6 +28113,8 @@ public Builder clearPartKey() { } /** * repeated string part_key = 2; + * @param value The bytes of the partKey to add. + * @return This builder for chaining. */ public Builder addPartKeyBytes( com.google.protobuf.ByteString value) { @@ -24527,35 +28127,40 @@ public Builder addPartKeyBytes( return this; } - // repeated string part_val = 3; private com.google.protobuf.LazyStringList partVal_ = com.google.protobuf.LazyStringArrayList.EMPTY; private void ensurePartValIsMutable() { - if (!((bitField0_ & 0x00000004) == 0x00000004)) { + if (!((bitField0_ & 0x00000004) != 0)) { partVal_ = new com.google.protobuf.LazyStringArrayList(partVal_); bitField0_ |= 0x00000004; } } /** * repeated string part_val = 3; + * @return A list containing the partVal. */ - public java.util.List + public com.google.protobuf.ProtocolStringList getPartValList() { - return java.util.Collections.unmodifiableList(partVal_); + return partVal_.getUnmodifiableView(); } /** * repeated string part_val = 3; + * @return The count of partVal. */ public int getPartValCount() { return partVal_.size(); } /** * repeated string part_val = 3; + * @param index The index of the element to return. + * @return The partVal at the given index. */ public java.lang.String getPartVal(int index) { return partVal_.get(index); } /** * repeated string part_val = 3; + * @param index The index of the value to return. + * @return The bytes of the partVal at the given index. */ public com.google.protobuf.ByteString getPartValBytes(int index) { @@ -24563,6 +28168,9 @@ public java.lang.String getPartVal(int index) { } /** * repeated string part_val = 3; + * @param index The index to set the value at. + * @param value The partVal to set. + * @return This builder for chaining. */ public Builder setPartVal( int index, java.lang.String value) { @@ -24576,6 +28184,8 @@ public Builder setPartVal( } /** * repeated string part_val = 3; + * @param value The partVal to add. + * @return This builder for chaining. */ public Builder addPartVal( java.lang.String value) { @@ -24589,16 +28199,20 @@ public Builder addPartVal( } /** * repeated string part_val = 3; + * @param values The partVal to add. + * @return This builder for chaining. */ public Builder addAllPartVal( java.lang.Iterable values) { ensurePartValIsMutable(); - super.addAll(values, partVal_); + com.google.protobuf.AbstractMessageLite.Builder.addAll( + values, partVal_); onChanged(); return this; } /** * repeated string part_val = 3; + * @return This builder for chaining. */ public Builder clearPartVal() { partVal_ = com.google.protobuf.LazyStringArrayList.EMPTY; @@ -24608,6 +28222,8 @@ public Builder clearPartVal() { } /** * repeated string part_val = 3; + * @param value The bytes of the partVal to add. + * @return This builder for chaining. */ public Builder addPartValBytes( com.google.protobuf.ByteString value) { @@ -24619,28 +28235,71 @@ public Builder addPartValBytes( onChanged(); return this; } + @java.lang.Override + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:TableProto) } + // @@protoc_insertion_point(class_scope:TableProto) + private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TableProto DEFAULT_INSTANCE; static { - defaultInstance = new TableProto(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TableProto(); + } + + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TableProto getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser + PARSER = new com.google.protobuf.AbstractParser() { + @java.lang.Override + public TableProto parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new TableProto(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.TableProto getDefaultInstanceForType() { + return DEFAULT_INSTANCE; } - // @@protoc_insertion_point(class_scope:TableProto) } - public interface EvictEntityResponseProtoOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface EvictEntityResponseProtoOrBuilder extends + // @@protoc_insertion_point(interface_extends:EvictEntityResponseProto) + com.google.protobuf.MessageOrBuilder { - // required int64 evicted_bytes = 1; /** * required int64 evicted_bytes = 1; + * @return Whether the evictedBytes field is set. */ boolean hasEvictedBytes(); /** * required int64 evicted_bytes = 1; + * @return The evictedBytes. */ long getEvictedBytes(); } @@ -24648,35 +28307,37 @@ public interface EvictEntityResponseProtoOrBuilder * Protobuf type {@code EvictEntityResponseProto} */ public static final class EvictEntityResponseProto extends - com.google.protobuf.GeneratedMessage - implements EvictEntityResponseProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:EvictEntityResponseProto) + EvictEntityResponseProtoOrBuilder { + private static final long serialVersionUID = 0L; // Use EvictEntityResponseProto.newBuilder() to construct. - private EvictEntityResponseProto(com.google.protobuf.GeneratedMessage.Builder builder) { + private EvictEntityResponseProto(com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private EvictEntityResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final EvictEntityResponseProto defaultInstance; - public static EvictEntityResponseProto getDefaultInstance() { - return defaultInstance; + private EvictEntityResponseProto() { } - public EvictEntityResponseProto getDefaultInstanceForType() { - return defaultInstance; + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance( + UnusedPrivateParameter unused) { + return new EvictEntityResponseProto(); } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private EvictEntityResponseProto( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); @@ -24688,25 +28349,25 @@ private EvictEntityResponseProto( case 0: done = true; break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } case 8: { bitField0_ |= 0x00000001; evictedBytes_ = input.readInt64(); break; } + default: { + if (!parseUnknownField( + input, unknownFields, extensionRegistry, tag)) { + done = true; + } + break; + } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); @@ -24717,52 +28378,40 @@ private EvictEntityResponseProto( return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_EvictEntityResponseProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_EvictEntityResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityResponseProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityResponseProto.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public EvictEntityResponseProto parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new EvictEntityResponseProto(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - private int bitField0_; - // required int64 evicted_bytes = 1; public static final int EVICTED_BYTES_FIELD_NUMBER = 1; private long evictedBytes_; /** * required int64 evicted_bytes = 1; + * @return Whether the evictedBytes field is set. */ + @java.lang.Override public boolean hasEvictedBytes() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * required int64 evicted_bytes = 1; + * @return The evictedBytes. */ + @java.lang.Override public long getEvictedBytes() { return evictedBytes_; } - private void initFields() { - evictedBytes_ = 0L; - } private byte memoizedIsInitialized = -1; + @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; if (!hasEvictedBytes()) { memoizedIsInitialized = 0; @@ -24772,37 +28421,30 @@ public final boolean isInitialized() { return true; } + @java.lang.Override public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); - if (((bitField0_ & 0x00000001) == 0x00000001)) { + if (((bitField0_ & 0x00000001) != 0)) { output.writeInt64(1, evictedBytes_); } - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; + @java.lang.Override public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { + if (((bitField0_ & 0x00000001) != 0)) { size += com.google.protobuf.CodedOutputStream .computeInt64Size(1, evictedBytes_); } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { @@ -24813,34 +28455,43 @@ public boolean equals(final java.lang.Object obj) { } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityResponseProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityResponseProto) obj; - boolean result = true; - result = result && (hasEvictedBytes() == other.hasEvictedBytes()); + if (hasEvictedBytes() != other.hasEvictedBytes()) return false; if (hasEvictedBytes()) { - result = result && (getEvictedBytes() - == other.getEvictedBytes()); + if (getEvictedBytes() + != other.getEvictedBytes()) return false; } - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; + if (!unknownFields.equals(other.unknownFields)) return false; + return true; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); + hash = (19 * hash) + getDescriptor().hashCode(); if (hasEvictedBytes()) { hash = (37 * hash) + EVICTED_BYTES_FIELD_NUMBER; - hash = (53 * hash) + hashLong(getEvictedBytes()); + hash = (53 * hash) + com.google.protobuf.Internal.hashLong( + getEvictedBytes()); } - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityResponseProto parseFrom( + java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityResponseProto parseFrom( + java.nio.ByteBuffer data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityResponseProto parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { @@ -24864,46 +28515,59 @@ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.Ev } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityResponseProto parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityResponseProto parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityResponseProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityResponseProto parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityResponseProto parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityResponseProto parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } + @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityResponseProto prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -24911,14 +28575,16 @@ protected Builder newBuilderForType( * Protobuf type {@code EvictEntityResponseProto} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityResponseProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder implements + // @@protoc_insertion_point(builder_implements:EvictEntityResponseProto) + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityResponseProtoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_EvictEntityResponseProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_EvictEntityResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -24931,18 +28597,16 @@ private Builder() { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { } } - private static Builder create() { - return new Builder(); - } - + @java.lang.Override public Builder clear() { super.clear(); evictedBytes_ = 0L; @@ -24950,19 +28614,18 @@ public Builder clear() { return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - + @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_EvictEntityResponseProto_descriptor; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityResponseProto getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityResponseProto.getDefaultInstance(); } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityResponseProto build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityResponseProto result = buildPartial(); if (!result.isInitialized()) { @@ -24971,19 +28634,53 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEnti return result; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityResponseProto buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityResponseProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityResponseProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { + if (((from_bitField0_ & 0x00000001) != 0)) { + result.evictedBytes_ = evictedBytes_; to_bitField0_ |= 0x00000001; } - result.evictedBytes_ = evictedBytes_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } + @java.lang.Override + public Builder clone() { + return super.clone(); + } + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.setField(field, value); + } + @java.lang.Override + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + @java.lang.Override + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.addRepeatedField(field, value); + } + @java.lang.Override public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityResponseProto) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityResponseProto)other); @@ -24998,18 +28695,20 @@ public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtoc if (other.hasEvictedBytes()) { setEvictedBytes(other.getEvictedBytes()); } - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } + @java.lang.Override public final boolean isInitialized() { if (!hasEvictedBytes()) { - return false; } return true; } + @java.lang.Override public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) @@ -25019,7 +28718,7 @@ public Builder mergeFrom( parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityResponseProto) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -25029,22 +28728,27 @@ public Builder mergeFrom( } private int bitField0_; - // required int64 evicted_bytes = 1; private long evictedBytes_ ; /** * required int64 evicted_bytes = 1; + * @return Whether the evictedBytes field is set. */ + @java.lang.Override public boolean hasEvictedBytes() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * required int64 evicted_bytes = 1; + * @return The evictedBytes. */ + @java.lang.Override public long getEvictedBytes() { return evictedBytes_; } /** * required int64 evicted_bytes = 1; + * @param value The evictedBytes to set. + * @return This builder for chaining. */ public Builder setEvictedBytes(long value) { bitField0_ |= 0x00000001; @@ -25054,6 +28758,7 @@ public Builder setEvictedBytes(long value) { } /** * required int64 evicted_bytes = 1; + * @return This builder for chaining. */ public Builder clearEvictedBytes() { bitField0_ = (bitField0_ & ~0x00000001); @@ -25061,54 +28766,98 @@ public Builder clearEvictedBytes() { onChanged(); return this; } + @java.lang.Override + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:EvictEntityResponseProto) } + // @@protoc_insertion_point(class_scope:EvictEntityResponseProto) + private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityResponseProto DEFAULT_INSTANCE; static { - defaultInstance = new EvictEntityResponseProto(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityResponseProto(); + } + + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityResponseProto getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser + PARSER = new com.google.protobuf.AbstractParser() { + @java.lang.Override + public EvictEntityResponseProto parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new EvictEntityResponseProto(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.EvictEntityResponseProto getDefaultInstanceForType() { + return DEFAULT_INSTANCE; } - // @@protoc_insertion_point(class_scope:EvictEntityResponseProto) } - public interface GetCacheContentRequestProtoOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface GetCacheContentRequestProtoOrBuilder extends + // @@protoc_insertion_point(interface_extends:GetCacheContentRequestProto) + com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code GetCacheContentRequestProto} */ public static final class GetCacheContentRequestProto extends - com.google.protobuf.GeneratedMessage - implements GetCacheContentRequestProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:GetCacheContentRequestProto) + GetCacheContentRequestProtoOrBuilder { + private static final long serialVersionUID = 0L; // Use GetCacheContentRequestProto.newBuilder() to construct. - private GetCacheContentRequestProto(com.google.protobuf.GeneratedMessage.Builder builder) { + private GetCacheContentRequestProto(com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private GetCacheContentRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final GetCacheContentRequestProto defaultInstance; - public static GetCacheContentRequestProto getDefaultInstance() { - return defaultInstance; + private GetCacheContentRequestProto() { } - public GetCacheContentRequestProto getDefaultInstanceForType() { - return defaultInstance; + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance( + UnusedPrivateParameter unused) { + return new GetCacheContentRequestProto(); } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private GetCacheContentRequestProto( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); try { @@ -25120,8 +28869,8 @@ private GetCacheContentRequestProto( done = true; break; default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { + if (!parseUnknownField( + input, unknownFields, extensionRegistry, tag)) { done = true; } break; @@ -25132,7 +28881,7 @@ private GetCacheContentRequestProto( throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); @@ -25143,63 +28892,42 @@ private GetCacheContentRequestProto( return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GetCacheContentRequestProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GetCacheContentRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentRequestProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentRequestProto.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public GetCacheContentRequestProto parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new GetCacheContentRequestProto(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - - private void initFields() { - } private byte memoizedIsInitialized = -1; + @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } + @java.lang.Override public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; + @java.lang.Override public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { @@ -25210,25 +28938,33 @@ public boolean equals(final java.lang.Object obj) { } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentRequestProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentRequestProto) obj; - boolean result = true; - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; + if (!unknownFields.equals(other.unknownFields)) return false; + return true; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (19 * hash) + getDescriptor().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentRequestProto parseFrom( + java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentRequestProto parseFrom( + java.nio.ByteBuffer data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentRequestProto parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { @@ -25252,46 +28988,59 @@ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.Ge } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentRequestProto parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentRequestProto parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentRequestProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentRequestProto parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentRequestProto parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentRequestProto parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } + @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentRequestProto prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -25299,14 +29048,16 @@ protected Builder newBuilderForType( * Protobuf type {@code GetCacheContentRequestProto} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentRequestProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder implements + // @@protoc_insertion_point(builder_implements:GetCacheContentRequestProto) + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentRequestProtoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GetCacheContentRequestProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GetCacheContentRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -25319,36 +29070,33 @@ private Builder() { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { } } - private static Builder create() { - return new Builder(); - } - + @java.lang.Override public Builder clear() { super.clear(); return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - + @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GetCacheContentRequestProto_descriptor; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentRequestProto getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentRequestProto.getDefaultInstance(); } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentRequestProto build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentRequestProto result = buildPartial(); if (!result.isInitialized()) { @@ -25357,12 +29105,46 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheC return result; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentRequestProto buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentRequestProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentRequestProto(this); onBuilt(); return result; } + @java.lang.Override + public Builder clone() { + return super.clone(); + } + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.setField(field, value); + } + @java.lang.Override + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + @java.lang.Override + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.addRepeatedField(field, value); + } + @java.lang.Override public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentRequestProto) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentRequestProto)other); @@ -25374,14 +29156,17 @@ public Builder mergeFrom(com.google.protobuf.Message other) { public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentRequestProto other) { if (other == org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentRequestProto.getDefaultInstance()) return this; - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } + @java.lang.Override public final boolean isInitialized() { return true; } + @java.lang.Override public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) @@ -25391,7 +29176,7 @@ public Builder mergeFrom( parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentRequestProto) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -25399,28 +29184,71 @@ public Builder mergeFrom( } return this; } + @java.lang.Override + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:GetCacheContentRequestProto) } + // @@protoc_insertion_point(class_scope:GetCacheContentRequestProto) + private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentRequestProto DEFAULT_INSTANCE; static { - defaultInstance = new GetCacheContentRequestProto(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentRequestProto(); + } + + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentRequestProto getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser + PARSER = new com.google.protobuf.AbstractParser() { + @java.lang.Override + public GetCacheContentRequestProto parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new GetCacheContentRequestProto(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentRequestProto getDefaultInstanceForType() { + return DEFAULT_INSTANCE; } - // @@protoc_insertion_point(class_scope:GetCacheContentRequestProto) } - public interface GetCacheContentResponseProtoOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface GetCacheContentResponseProtoOrBuilder extends + // @@protoc_insertion_point(interface_extends:GetCacheContentResponseProto) + com.google.protobuf.MessageOrBuilder { - // optional .CacheEntryList result = 1; /** * optional .CacheEntryList result = 1; + * @return Whether the result field is set. */ boolean hasResult(); /** * optional .CacheEntryList result = 1; + * @return The result. */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList getResult(); /** @@ -25432,35 +29260,37 @@ public interface GetCacheContentResponseProtoOrBuilder * Protobuf type {@code GetCacheContentResponseProto} */ public static final class GetCacheContentResponseProto extends - com.google.protobuf.GeneratedMessage - implements GetCacheContentResponseProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:GetCacheContentResponseProto) + GetCacheContentResponseProtoOrBuilder { + private static final long serialVersionUID = 0L; // Use GetCacheContentResponseProto.newBuilder() to construct. - private GetCacheContentResponseProto(com.google.protobuf.GeneratedMessage.Builder builder) { + private GetCacheContentResponseProto(com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private GetCacheContentResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final GetCacheContentResponseProto defaultInstance; - public static GetCacheContentResponseProto getDefaultInstance() { - return defaultInstance; + private GetCacheContentResponseProto() { } - public GetCacheContentResponseProto getDefaultInstanceForType() { - return defaultInstance; + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance( + UnusedPrivateParameter unused) { + return new GetCacheContentResponseProto(); } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private GetCacheContentResponseProto( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); @@ -25472,16 +29302,9 @@ private GetCacheContentResponseProto( case 0: done = true; break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } case 10: { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList.Builder subBuilder = null; - if (((bitField0_ & 0x00000001) == 0x00000001)) { + if (((bitField0_ & 0x00000001) != 0)) { subBuilder = result_.toBuilder(); } result_ = input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList.PARSER, extensionRegistry); @@ -25492,13 +29315,20 @@ private GetCacheContentResponseProto( bitField0_ |= 0x00000001; break; } + default: { + if (!parseUnknownField( + input, unknownFields, extensionRegistry, tag)) { + done = true; + } + break; + } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); @@ -25509,94 +29339,76 @@ private GetCacheContentResponseProto( return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GetCacheContentResponseProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GetCacheContentResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentResponseProto.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentResponseProto.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public GetCacheContentResponseProto parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new GetCacheContentResponseProto(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - private int bitField0_; - // optional .CacheEntryList result = 1; public static final int RESULT_FIELD_NUMBER = 1; private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList result_; /** * optional .CacheEntryList result = 1; + * @return Whether the result field is set. */ + @java.lang.Override public boolean hasResult() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional .CacheEntryList result = 1; + * @return The result. */ + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList getResult() { - return result_; + return result_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList.getDefaultInstance() : result_; } /** * optional .CacheEntryList result = 1; */ + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryListOrBuilder getResultOrBuilder() { - return result_; + return result_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList.getDefaultInstance() : result_; } - private void initFields() { - result_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList.getDefaultInstance(); - } private byte memoizedIsInitialized = -1; + @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } + @java.lang.Override public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); - if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeMessage(1, result_); + if (((bitField0_ & 0x00000001) != 0)) { + output.writeMessage(1, getResult()); } - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; + @java.lang.Override public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { + if (((bitField0_ & 0x00000001) != 0)) { size += com.google.protobuf.CodedOutputStream - .computeMessageSize(1, result_); + .computeMessageSize(1, getResult()); } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { @@ -25607,34 +29419,42 @@ public boolean equals(final java.lang.Object obj) { } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentResponseProto other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentResponseProto) obj; - boolean result = true; - result = result && (hasResult() == other.hasResult()); + if (hasResult() != other.hasResult()) return false; if (hasResult()) { - result = result && getResult() - .equals(other.getResult()); + if (!getResult() + .equals(other.getResult())) return false; } - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; + if (!unknownFields.equals(other.unknownFields)) return false; + return true; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); + hash = (19 * hash) + getDescriptor().hashCode(); if (hasResult()) { hash = (37 * hash) + RESULT_FIELD_NUMBER; hash = (53 * hash) + getResult().hashCode(); } - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentResponseProto parseFrom( + java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentResponseProto parseFrom( + java.nio.ByteBuffer data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentResponseProto parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { @@ -25658,46 +29478,59 @@ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.Ge } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentResponseProto parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentResponseProto parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentResponseProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentResponseProto parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentResponseProto parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentResponseProto parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } + @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentResponseProto prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -25705,14 +29538,16 @@ protected Builder newBuilderForType( * Protobuf type {@code GetCacheContentResponseProto} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentResponseProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder implements + // @@protoc_insertion_point(builder_implements:GetCacheContentResponseProto) + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentResponseProtoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GetCacheContentResponseProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GetCacheContentResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -25725,23 +29560,21 @@ private Builder() { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { getResultFieldBuilder(); } } - private static Builder create() { - return new Builder(); - } - + @java.lang.Override public Builder clear() { super.clear(); if (resultBuilder_ == null) { - result_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList.getDefaultInstance(); + result_ = null; } else { resultBuilder_.clear(); } @@ -25749,19 +29582,18 @@ public Builder clear() { return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - + @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_GetCacheContentResponseProto_descriptor; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentResponseProto getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentResponseProto.getDefaultInstance(); } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentResponseProto build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentResponseProto result = buildPartial(); if (!result.isInitialized()) { @@ -25770,23 +29602,57 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheC return result; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentResponseProto buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentResponseProto result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentResponseProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { + if (((from_bitField0_ & 0x00000001) != 0)) { + if (resultBuilder_ == null) { + result.result_ = result_; + } else { + result.result_ = resultBuilder_.build(); + } to_bitField0_ |= 0x00000001; } - if (resultBuilder_ == null) { - result.result_ = result_; - } else { - result.result_ = resultBuilder_.build(); - } result.bitField0_ = to_bitField0_; onBuilt(); return result; } + @java.lang.Override + public Builder clone() { + return super.clone(); + } + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.setField(field, value); + } + @java.lang.Override + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + @java.lang.Override + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.addRepeatedField(field, value); + } + @java.lang.Override public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentResponseProto) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentResponseProto)other); @@ -25801,14 +29667,17 @@ public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtoc if (other.hasResult()) { mergeResult(other.getResult()); } - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } + @java.lang.Override public final boolean isInitialized() { return true; } + @java.lang.Override public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) @@ -25818,7 +29687,7 @@ public Builder mergeFrom( parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentResponseProto) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -25828,22 +29697,23 @@ public Builder mergeFrom( } private int bitField0_; - // optional .CacheEntryList result = 1; - private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList result_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList.getDefaultInstance(); - private com.google.protobuf.SingleFieldBuilder< + private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList result_; + private com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryListOrBuilder> resultBuilder_; /** * optional .CacheEntryList result = 1; + * @return Whether the result field is set. */ public boolean hasResult() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional .CacheEntryList result = 1; + * @return The result. */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList getResult() { if (resultBuilder_ == null) { - return result_; + return result_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList.getDefaultInstance() : result_; } else { return resultBuilder_.getMessage(); } @@ -25883,7 +29753,8 @@ public Builder setResult( */ public Builder mergeResult(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList value) { if (resultBuilder_ == null) { - if (((bitField0_ & 0x00000001) == 0x00000001) && + if (((bitField0_ & 0x00000001) != 0) && + result_ != null && result_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList.getDefaultInstance()) { result_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList.newBuilder(result_).mergeFrom(value).buildPartial(); @@ -25902,7 +29773,7 @@ public Builder mergeResult(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProt */ public Builder clearResult() { if (resultBuilder_ == null) { - result_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList.getDefaultInstance(); + result_ = null; onChanged(); } else { resultBuilder_.clear(); @@ -25925,41 +29796,83 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntr if (resultBuilder_ != null) { return resultBuilder_.getMessageOrBuilder(); } else { - return result_; + return result_ == null ? + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList.getDefaultInstance() : result_; } } /** * optional .CacheEntryList result = 1; */ - private com.google.protobuf.SingleFieldBuilder< + private com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryListOrBuilder> getResultFieldBuilder() { if (resultBuilder_ == null) { - resultBuilder_ = new com.google.protobuf.SingleFieldBuilder< + resultBuilder_ = new com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryListOrBuilder>( - result_, + getResult(), getParentForChildren(), isClean()); result_ = null; } return resultBuilder_; } + @java.lang.Override + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:GetCacheContentResponseProto) } + // @@protoc_insertion_point(class_scope:GetCacheContentResponseProto) + private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentResponseProto DEFAULT_INSTANCE; static { - defaultInstance = new GetCacheContentResponseProto(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentResponseProto(); + } + + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentResponseProto getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser + PARSER = new com.google.protobuf.AbstractParser() { + @java.lang.Override + public GetCacheContentResponseProto parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new GetCacheContentResponseProto(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheContentResponseProto getDefaultInstanceForType() { + return DEFAULT_INSTANCE; } - // @@protoc_insertion_point(class_scope:GetCacheContentResponseProto) } - public interface CacheEntryListOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface CacheEntryListOrBuilder extends + // @@protoc_insertion_point(interface_extends:CacheEntryList) + com.google.protobuf.MessageOrBuilder { - // repeated .CacheEntry entries = 1; /** * repeated .CacheEntry entries = 1; */ @@ -25988,35 +29901,38 @@ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryOrBuil * Protobuf type {@code CacheEntryList} */ public static final class CacheEntryList extends - com.google.protobuf.GeneratedMessage - implements CacheEntryListOrBuilder { + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:CacheEntryList) + CacheEntryListOrBuilder { + private static final long serialVersionUID = 0L; // Use CacheEntryList.newBuilder() to construct. - private CacheEntryList(com.google.protobuf.GeneratedMessage.Builder builder) { + private CacheEntryList(com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private CacheEntryList(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final CacheEntryList defaultInstance; - public static CacheEntryList getDefaultInstance() { - return defaultInstance; + private CacheEntryList() { + entries_ = java.util.Collections.emptyList(); } - public CacheEntryList getDefaultInstanceForType() { - return defaultInstance; + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance( + UnusedPrivateParameter unused) { + return new CacheEntryList(); } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private CacheEntryList( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); @@ -26028,19 +29944,20 @@ private CacheEntryList( case 0: done = true; break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } case 10: { - if (!((mutable_bitField0_ & 0x00000001) == 0x00000001)) { + if (!((mutable_bitField0_ & 0x00000001) != 0)) { entries_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000001; } - entries_.add(input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntry.PARSER, extensionRegistry)); + entries_.add( + input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntry.PARSER, extensionRegistry)); + break; + } + default: { + if (!parseUnknownField( + input, unknownFields, extensionRegistry, tag)) { + done = true; + } break; } } @@ -26049,9 +29966,9 @@ private CacheEntryList( throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { - if (((mutable_bitField0_ & 0x00000001) == 0x00000001)) { + if (((mutable_bitField0_ & 0x00000001) != 0)) { entries_ = java.util.Collections.unmodifiableList(entries_); } this.unknownFields = unknownFields.build(); @@ -26063,40 +29980,27 @@ private CacheEntryList( return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_CacheEntryList_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_CacheEntryList_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public CacheEntryList parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new CacheEntryList(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - - // repeated .CacheEntry entries = 1; public static final int ENTRIES_FIELD_NUMBER = 1; private java.util.List entries_; /** * repeated .CacheEntry entries = 1; */ + @java.lang.Override public java.util.List getEntriesList() { return entries_; } /** * repeated .CacheEntry entries = 1; */ + @java.lang.Override public java.util.List getEntriesOrBuilderList() { return entries_; @@ -26104,47 +30008,49 @@ public java.util.Listrepeated .CacheEntry entries = 1; */ + @java.lang.Override public int getEntriesCount() { return entries_.size(); } /** * repeated .CacheEntry entries = 1; */ + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntry getEntries(int index) { return entries_.get(index); } /** * repeated .CacheEntry entries = 1; */ + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryOrBuilder getEntriesOrBuilder( int index) { return entries_.get(index); } - private void initFields() { - entries_ = java.util.Collections.emptyList(); - } private byte memoizedIsInitialized = -1; + @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } + @java.lang.Override public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); for (int i = 0; i < entries_.size(); i++) { output.writeMessage(1, entries_.get(i)); } - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; + @java.lang.Override public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; @@ -26152,18 +30058,11 @@ public int getSerializedSize() { size += com.google.protobuf.CodedOutputStream .computeMessageSize(1, entries_.get(i)); } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { @@ -26174,31 +30073,39 @@ public boolean equals(final java.lang.Object obj) { } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList) obj; - boolean result = true; - result = result && getEntriesList() - .equals(other.getEntriesList()); - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; + if (!getEntriesList() + .equals(other.getEntriesList())) return false; + if (!unknownFields.equals(other.unknownFields)) return false; + return true; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); + hash = (19 * hash) + getDescriptor().hashCode(); if (getEntriesCount() > 0) { hash = (37 * hash) + ENTRIES_FIELD_NUMBER; hash = (53 * hash) + getEntriesList().hashCode(); } - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList parseFrom( + java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList parseFrom( + java.nio.ByteBuffer data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { @@ -26222,46 +30129,59 @@ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.Ca } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } + @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -26269,14 +30189,16 @@ protected Builder newBuilderForType( * Protobuf type {@code CacheEntryList} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryListOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder implements + // @@protoc_insertion_point(builder_implements:CacheEntryList) + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryListOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_CacheEntryList_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_CacheEntryList_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -26289,19 +30211,17 @@ private Builder() { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { getEntriesFieldBuilder(); } } - private static Builder create() { - return new Builder(); - } - + @java.lang.Override public Builder clear() { super.clear(); if (entriesBuilder_ == null) { @@ -26313,19 +30233,18 @@ public Builder clear() { return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - + @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_CacheEntryList_descriptor; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList.getDefaultInstance(); } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList result = buildPartial(); if (!result.isInitialized()) { @@ -26334,11 +30253,12 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntr return result; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList(this); int from_bitField0_ = bitField0_; if (entriesBuilder_ == null) { - if (((bitField0_ & 0x00000001) == 0x00000001)) { + if (((bitField0_ & 0x00000001) != 0)) { entries_ = java.util.Collections.unmodifiableList(entries_); bitField0_ = (bitField0_ & ~0x00000001); } @@ -26350,6 +30270,39 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntr return result; } + @java.lang.Override + public Builder clone() { + return super.clone(); + } + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.setField(field, value); + } + @java.lang.Override + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + @java.lang.Override + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.addRepeatedField(field, value); + } + @java.lang.Override public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList)other); @@ -26380,21 +30333,24 @@ public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtoc entries_ = other.entries_; bitField0_ = (bitField0_ & ~0x00000001); entriesBuilder_ = - com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? + com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getEntriesFieldBuilder() : null; } else { entriesBuilder_.addAllMessages(other.entries_); } } } - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } + @java.lang.Override public final boolean isInitialized() { return true; } + @java.lang.Override public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) @@ -26404,7 +30360,7 @@ public Builder mergeFrom( parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -26414,17 +30370,16 @@ public Builder mergeFrom( } private int bitField0_; - // repeated .CacheEntry entries = 1; private java.util.List entries_ = java.util.Collections.emptyList(); private void ensureEntriesIsMutable() { - if (!((bitField0_ & 0x00000001) == 0x00000001)) { + if (!((bitField0_ & 0x00000001) != 0)) { entries_ = new java.util.ArrayList(entries_); bitField0_ |= 0x00000001; } } - private com.google.protobuf.RepeatedFieldBuilder< + private com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntry, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntry.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryOrBuilder> entriesBuilder_; /** @@ -26556,7 +30511,8 @@ public Builder addAllEntries( java.lang.Iterable values) { if (entriesBuilder_ == null) { ensureEntriesIsMutable(); - super.addAll(values, entries_); + com.google.protobuf.AbstractMessageLite.Builder.addAll( + values, entries_); onChanged(); } else { entriesBuilder_.addAllMessages(values); @@ -26639,67 +30595,113 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntr getEntriesBuilderList() { return getEntriesFieldBuilder().getBuilderList(); } - private com.google.protobuf.RepeatedFieldBuilder< + private com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntry, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntry.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryOrBuilder> getEntriesFieldBuilder() { if (entriesBuilder_ == null) { - entriesBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< + entriesBuilder_ = new com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntry, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntry.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryOrBuilder>( entries_, - ((bitField0_ & 0x00000001) == 0x00000001), + ((bitField0_ & 0x00000001) != 0), getParentForChildren(), isClean()); entries_ = null; } return entriesBuilder_; } + @java.lang.Override + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:CacheEntryList) } + // @@protoc_insertion_point(class_scope:CacheEntryList) + private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList DEFAULT_INSTANCE; static { - defaultInstance = new CacheEntryList(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList(); + } + + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser + PARSER = new com.google.protobuf.AbstractParser() { + @java.lang.Override + public CacheEntryList parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new CacheEntryList(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryList getDefaultInstanceForType() { + return DEFAULT_INSTANCE; } - // @@protoc_insertion_point(class_scope:CacheEntryList) } - public interface CacheEntryOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface CacheEntryOrBuilder extends + // @@protoc_insertion_point(interface_extends:CacheEntry) + com.google.protobuf.MessageOrBuilder { - // optional bytes file_key = 1; /** * optional bytes file_key = 1; + * @return Whether the fileKey field is set. */ boolean hasFileKey(); /** * optional bytes file_key = 1; + * @return The fileKey. */ com.google.protobuf.ByteString getFileKey(); - // optional string file_path = 2; /** * optional string file_path = 2; + * @return Whether the filePath field is set. */ boolean hasFilePath(); /** * optional string file_path = 2; + * @return The filePath. */ java.lang.String getFilePath(); /** * optional string file_path = 2; + * @return The bytes for filePath. */ com.google.protobuf.ByteString getFilePathBytes(); - // optional .CacheTag cache_tag = 3; /** * optional .CacheTag cache_tag = 3; + * @return Whether the cacheTag field is set. */ boolean hasCacheTag(); /** * optional .CacheTag cache_tag = 3; + * @return The cacheTag. */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag getCacheTag(); /** @@ -26707,7 +30709,6 @@ public interface CacheEntryOrBuilder */ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTagOrBuilder getCacheTagOrBuilder(); - // repeated .CacheEntryRange ranges = 4; /** * repeated .CacheEntryRange ranges = 4; */ @@ -26736,35 +30737,40 @@ org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryRangeO * Protobuf type {@code CacheEntry} */ public static final class CacheEntry extends - com.google.protobuf.GeneratedMessage - implements CacheEntryOrBuilder { + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:CacheEntry) + CacheEntryOrBuilder { + private static final long serialVersionUID = 0L; // Use CacheEntry.newBuilder() to construct. - private CacheEntry(com.google.protobuf.GeneratedMessage.Builder builder) { + private CacheEntry(com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private CacheEntry(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final CacheEntry defaultInstance; - public static CacheEntry getDefaultInstance() { - return defaultInstance; + private CacheEntry() { + fileKey_ = com.google.protobuf.ByteString.EMPTY; + filePath_ = ""; + ranges_ = java.util.Collections.emptyList(); } - public CacheEntry getDefaultInstanceForType() { - return defaultInstance; + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance( + UnusedPrivateParameter unused) { + return new CacheEntry(); } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private CacheEntry( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); @@ -26776,26 +30782,20 @@ private CacheEntry( case 0: done = true; break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } case 10: { bitField0_ |= 0x00000001; fileKey_ = input.readBytes(); break; } case 18: { + com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000002; - filePath_ = input.readBytes(); + filePath_ = bs; break; } case 26: { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag.Builder subBuilder = null; - if (((bitField0_ & 0x00000004) == 0x00000004)) { + if (((bitField0_ & 0x00000004) != 0)) { subBuilder = cacheTag_.toBuilder(); } cacheTag_ = input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag.PARSER, extensionRegistry); @@ -26807,11 +30807,19 @@ private CacheEntry( break; } case 34: { - if (!((mutable_bitField0_ & 0x00000008) == 0x00000008)) { + if (!((mutable_bitField0_ & 0x00000008) != 0)) { ranges_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000008; } - ranges_.add(input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryRange.PARSER, extensionRegistry)); + ranges_.add( + input.readMessage(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryRange.PARSER, extensionRegistry)); + break; + } + default: { + if (!parseUnknownField( + input, unknownFields, extensionRegistry, tag)) { + done = true; + } break; } } @@ -26820,9 +30828,9 @@ private CacheEntry( throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { - if (((mutable_bitField0_ & 0x00000008) == 0x00000008)) { + if (((mutable_bitField0_ & 0x00000008) != 0)) { ranges_ = java.util.Collections.unmodifiableList(ranges_); } this.unknownFields = unknownFields.build(); @@ -26834,57 +30842,49 @@ private CacheEntry( return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_CacheEntry_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_CacheEntry_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntry.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntry.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public CacheEntry parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new CacheEntry(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - private int bitField0_; - // optional bytes file_key = 1; public static final int FILE_KEY_FIELD_NUMBER = 1; private com.google.protobuf.ByteString fileKey_; /** * optional bytes file_key = 1; + * @return Whether the fileKey field is set. */ + @java.lang.Override public boolean hasFileKey() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional bytes file_key = 1; + * @return The fileKey. */ + @java.lang.Override public com.google.protobuf.ByteString getFileKey() { return fileKey_; } - // optional string file_path = 2; public static final int FILE_PATH_FIELD_NUMBER = 2; - private java.lang.Object filePath_; + private volatile java.lang.Object filePath_; /** * optional string file_path = 2; + * @return Whether the filePath field is set. */ + @java.lang.Override public boolean hasFilePath() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * optional string file_path = 2; + * @return The filePath. */ + @java.lang.Override public java.lang.String getFilePath() { java.lang.Object ref = filePath_; if (ref instanceof java.lang.String) { @@ -26901,7 +30901,9 @@ public java.lang.String getFilePath() { } /** * optional string file_path = 2; + * @return The bytes for filePath. */ + @java.lang.Override public com.google.protobuf.ByteString getFilePathBytes() { java.lang.Object ref = filePath_; @@ -26916,40 +30918,45 @@ public java.lang.String getFilePath() { } } - // optional .CacheTag cache_tag = 3; public static final int CACHE_TAG_FIELD_NUMBER = 3; private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag cacheTag_; /** * optional .CacheTag cache_tag = 3; + * @return Whether the cacheTag field is set. */ + @java.lang.Override public boolean hasCacheTag() { - return ((bitField0_ & 0x00000004) == 0x00000004); + return ((bitField0_ & 0x00000004) != 0); } /** * optional .CacheTag cache_tag = 3; + * @return The cacheTag. */ + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag getCacheTag() { - return cacheTag_; + return cacheTag_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag.getDefaultInstance() : cacheTag_; } /** * optional .CacheTag cache_tag = 3; */ + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTagOrBuilder getCacheTagOrBuilder() { - return cacheTag_; + return cacheTag_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag.getDefaultInstance() : cacheTag_; } - // repeated .CacheEntryRange ranges = 4; public static final int RANGES_FIELD_NUMBER = 4; private java.util.List ranges_; /** * repeated .CacheEntryRange ranges = 4; */ + @java.lang.Override public java.util.List getRangesList() { return ranges_; } /** * repeated .CacheEntryRange ranges = 4; */ + @java.lang.Override public java.util.List getRangesOrBuilderList() { return ranges_; @@ -26957,90 +30964,81 @@ public java.util.Listrepeated .CacheEntryRange ranges = 4; */ + @java.lang.Override public int getRangesCount() { return ranges_.size(); } /** * repeated .CacheEntryRange ranges = 4; */ + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryRange getRanges(int index) { return ranges_.get(index); } /** * repeated .CacheEntryRange ranges = 4; */ + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryRangeOrBuilder getRangesOrBuilder( int index) { return ranges_.get(index); } - private void initFields() { - fileKey_ = com.google.protobuf.ByteString.EMPTY; - filePath_ = ""; - cacheTag_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag.getDefaultInstance(); - ranges_ = java.util.Collections.emptyList(); - } private byte memoizedIsInitialized = -1; + @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } + @java.lang.Override public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); - if (((bitField0_ & 0x00000001) == 0x00000001)) { + if (((bitField0_ & 0x00000001) != 0)) { output.writeBytes(1, fileKey_); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - output.writeBytes(2, getFilePathBytes()); + if (((bitField0_ & 0x00000002) != 0)) { + com.google.protobuf.GeneratedMessageV3.writeString(output, 2, filePath_); } - if (((bitField0_ & 0x00000004) == 0x00000004)) { - output.writeMessage(3, cacheTag_); + if (((bitField0_ & 0x00000004) != 0)) { + output.writeMessage(3, getCacheTag()); } for (int i = 0; i < ranges_.size(); i++) { output.writeMessage(4, ranges_.get(i)); } - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; + @java.lang.Override public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { + if (((bitField0_ & 0x00000001) != 0)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(1, fileKey_); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(2, getFilePathBytes()); + if (((bitField0_ & 0x00000002) != 0)) { + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(2, filePath_); } - if (((bitField0_ & 0x00000004) == 0x00000004)) { + if (((bitField0_ & 0x00000004) != 0)) { size += com.google.protobuf.CodedOutputStream - .computeMessageSize(3, cacheTag_); + .computeMessageSize(3, getCacheTag()); } for (int i = 0; i < ranges_.size(); i++) { size += com.google.protobuf.CodedOutputStream .computeMessageSize(4, ranges_.get(i)); } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { @@ -27051,37 +31049,34 @@ public boolean equals(final java.lang.Object obj) { } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntry other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntry) obj; - boolean result = true; - result = result && (hasFileKey() == other.hasFileKey()); + if (hasFileKey() != other.hasFileKey()) return false; if (hasFileKey()) { - result = result && getFileKey() - .equals(other.getFileKey()); + if (!getFileKey() + .equals(other.getFileKey())) return false; } - result = result && (hasFilePath() == other.hasFilePath()); + if (hasFilePath() != other.hasFilePath()) return false; if (hasFilePath()) { - result = result && getFilePath() - .equals(other.getFilePath()); + if (!getFilePath() + .equals(other.getFilePath())) return false; } - result = result && (hasCacheTag() == other.hasCacheTag()); + if (hasCacheTag() != other.hasCacheTag()) return false; if (hasCacheTag()) { - result = result && getCacheTag() - .equals(other.getCacheTag()); + if (!getCacheTag() + .equals(other.getCacheTag())) return false; } - result = result && getRangesList() - .equals(other.getRangesList()); - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; + if (!getRangesList() + .equals(other.getRangesList())) return false; + if (!unknownFields.equals(other.unknownFields)) return false; + return true; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); + hash = (19 * hash) + getDescriptor().hashCode(); if (hasFileKey()) { hash = (37 * hash) + FILE_KEY_FIELD_NUMBER; hash = (53 * hash) + getFileKey().hashCode(); @@ -27098,11 +31093,22 @@ public int hashCode() { hash = (37 * hash) + RANGES_FIELD_NUMBER; hash = (53 * hash) + getRangesList().hashCode(); } - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntry parseFrom( + java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntry parseFrom( + java.nio.ByteBuffer data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntry parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { @@ -27126,46 +31132,59 @@ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.Ca } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntry parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntry parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntry parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntry parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntry parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntry parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } + @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntry prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -27173,14 +31192,16 @@ protected Builder newBuilderForType( * Protobuf type {@code CacheEntry} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder implements + // @@protoc_insertion_point(builder_implements:CacheEntry) + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_CacheEntry_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_CacheEntry_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -27193,20 +31214,18 @@ private Builder() { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { getCacheTagFieldBuilder(); getRangesFieldBuilder(); } } - private static Builder create() { - return new Builder(); - } - + @java.lang.Override public Builder clear() { super.clear(); fileKey_ = com.google.protobuf.ByteString.EMPTY; @@ -27214,7 +31233,7 @@ public Builder clear() { filePath_ = ""; bitField0_ = (bitField0_ & ~0x00000002); if (cacheTagBuilder_ == null) { - cacheTag_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag.getDefaultInstance(); + cacheTag_ = null; } else { cacheTagBuilder_.clear(); } @@ -27228,19 +31247,18 @@ public Builder clear() { return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - + @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_CacheEntry_descriptor; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntry getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntry.getDefaultInstance(); } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntry build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntry result = buildPartial(); if (!result.isInitialized()) { @@ -27249,28 +31267,29 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntr return result; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntry buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntry result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntry(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { + if (((from_bitField0_ & 0x00000001) != 0)) { to_bitField0_ |= 0x00000001; } result.fileKey_ = fileKey_; - if (((from_bitField0_ & 0x00000002) == 0x00000002)) { + if (((from_bitField0_ & 0x00000002) != 0)) { to_bitField0_ |= 0x00000002; } result.filePath_ = filePath_; - if (((from_bitField0_ & 0x00000004) == 0x00000004)) { + if (((from_bitField0_ & 0x00000004) != 0)) { + if (cacheTagBuilder_ == null) { + result.cacheTag_ = cacheTag_; + } else { + result.cacheTag_ = cacheTagBuilder_.build(); + } to_bitField0_ |= 0x00000004; } - if (cacheTagBuilder_ == null) { - result.cacheTag_ = cacheTag_; - } else { - result.cacheTag_ = cacheTagBuilder_.build(); - } if (rangesBuilder_ == null) { - if (((bitField0_ & 0x00000008) == 0x00000008)) { + if (((bitField0_ & 0x00000008) != 0)) { ranges_ = java.util.Collections.unmodifiableList(ranges_); bitField0_ = (bitField0_ & ~0x00000008); } @@ -27283,6 +31302,39 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntr return result; } + @java.lang.Override + public Builder clone() { + return super.clone(); + } + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.setField(field, value); + } + @java.lang.Override + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + @java.lang.Override + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.addRepeatedField(field, value); + } + @java.lang.Override public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntry) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntry)other); @@ -27324,21 +31376,24 @@ public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtoc ranges_ = other.ranges_; bitField0_ = (bitField0_ & ~0x00000008); rangesBuilder_ = - com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? + com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getRangesFieldBuilder() : null; } else { rangesBuilder_.addAllMessages(other.ranges_); } } } - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } + @java.lang.Override public final boolean isInitialized() { return true; } + @java.lang.Override public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) @@ -27348,7 +31403,7 @@ public Builder mergeFrom( parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntry) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -27358,22 +31413,27 @@ public Builder mergeFrom( } private int bitField0_; - // optional bytes file_key = 1; private com.google.protobuf.ByteString fileKey_ = com.google.protobuf.ByteString.EMPTY; /** * optional bytes file_key = 1; + * @return Whether the fileKey field is set. */ + @java.lang.Override public boolean hasFileKey() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional bytes file_key = 1; + * @return The fileKey. */ + @java.lang.Override public com.google.protobuf.ByteString getFileKey() { return fileKey_; } /** * optional bytes file_key = 1; + * @param value The fileKey to set. + * @return This builder for chaining. */ public Builder setFileKey(com.google.protobuf.ByteString value) { if (value == null) { @@ -27386,6 +31446,7 @@ public Builder setFileKey(com.google.protobuf.ByteString value) { } /** * optional bytes file_key = 1; + * @return This builder for chaining. */ public Builder clearFileKey() { bitField0_ = (bitField0_ & ~0x00000001); @@ -27394,23 +31455,27 @@ public Builder clearFileKey() { return this; } - // optional string file_path = 2; private java.lang.Object filePath_ = ""; /** * optional string file_path = 2; + * @return Whether the filePath field is set. */ public boolean hasFilePath() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * optional string file_path = 2; + * @return The filePath. */ public java.lang.String getFilePath() { java.lang.Object ref = filePath_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - filePath_ = s; + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + filePath_ = s; + } return s; } else { return (java.lang.String) ref; @@ -27418,6 +31483,7 @@ public java.lang.String getFilePath() { } /** * optional string file_path = 2; + * @return The bytes for filePath. */ public com.google.protobuf.ByteString getFilePathBytes() { @@ -27434,6 +31500,8 @@ public java.lang.String getFilePath() { } /** * optional string file_path = 2; + * @param value The filePath to set. + * @return This builder for chaining. */ public Builder setFilePath( java.lang.String value) { @@ -27447,6 +31515,7 @@ public Builder setFilePath( } /** * optional string file_path = 2; + * @return This builder for chaining. */ public Builder clearFilePath() { bitField0_ = (bitField0_ & ~0x00000002); @@ -27456,6 +31525,8 @@ public Builder clearFilePath() { } /** * optional string file_path = 2; + * @param value The bytes for filePath to set. + * @return This builder for chaining. */ public Builder setFilePathBytes( com.google.protobuf.ByteString value) { @@ -27468,22 +31539,23 @@ public Builder setFilePathBytes( return this; } - // optional .CacheTag cache_tag = 3; - private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag cacheTag_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag.getDefaultInstance(); - private com.google.protobuf.SingleFieldBuilder< + private org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag cacheTag_; + private com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTagOrBuilder> cacheTagBuilder_; /** * optional .CacheTag cache_tag = 3; + * @return Whether the cacheTag field is set. */ public boolean hasCacheTag() { - return ((bitField0_ & 0x00000004) == 0x00000004); + return ((bitField0_ & 0x00000004) != 0); } /** * optional .CacheTag cache_tag = 3; + * @return The cacheTag. */ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag getCacheTag() { if (cacheTagBuilder_ == null) { - return cacheTag_; + return cacheTag_ == null ? org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag.getDefaultInstance() : cacheTag_; } else { return cacheTagBuilder_.getMessage(); } @@ -27523,7 +31595,8 @@ public Builder setCacheTag( */ public Builder mergeCacheTag(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag value) { if (cacheTagBuilder_ == null) { - if (((bitField0_ & 0x00000004) == 0x00000004) && + if (((bitField0_ & 0x00000004) != 0) && + cacheTag_ != null && cacheTag_ != org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag.getDefaultInstance()) { cacheTag_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag.newBuilder(cacheTag_).mergeFrom(value).buildPartial(); @@ -27542,7 +31615,7 @@ public Builder mergeCacheTag(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonPr */ public Builder clearCacheTag() { if (cacheTagBuilder_ == null) { - cacheTag_ = org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag.getDefaultInstance(); + cacheTag_ = null; onChanged(); } else { cacheTagBuilder_.clear(); @@ -27565,19 +31638,20 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTagO if (cacheTagBuilder_ != null) { return cacheTagBuilder_.getMessageOrBuilder(); } else { - return cacheTag_; + return cacheTag_ == null ? + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag.getDefaultInstance() : cacheTag_; } } /** * optional .CacheTag cache_tag = 3; */ - private com.google.protobuf.SingleFieldBuilder< + private com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTagOrBuilder> getCacheTagFieldBuilder() { if (cacheTagBuilder_ == null) { - cacheTagBuilder_ = new com.google.protobuf.SingleFieldBuilder< + cacheTagBuilder_ = new com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTagOrBuilder>( - cacheTag_, + getCacheTag(), getParentForChildren(), isClean()); cacheTag_ = null; @@ -27585,17 +31659,16 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTagO return cacheTagBuilder_; } - // repeated .CacheEntryRange ranges = 4; private java.util.List ranges_ = java.util.Collections.emptyList(); private void ensureRangesIsMutable() { - if (!((bitField0_ & 0x00000008) == 0x00000008)) { + if (!((bitField0_ & 0x00000008) != 0)) { ranges_ = new java.util.ArrayList(ranges_); bitField0_ |= 0x00000008; } } - private com.google.protobuf.RepeatedFieldBuilder< + private com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryRange, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryRange.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryRangeOrBuilder> rangesBuilder_; /** @@ -27727,7 +31800,8 @@ public Builder addAllRanges( java.lang.Iterable values) { if (rangesBuilder_ == null) { ensureRangesIsMutable(); - super.addAll(values, ranges_); + com.google.protobuf.AbstractMessageLite.Builder.addAll( + values, ranges_); onChanged(); } else { rangesBuilder_.addAllMessages(values); @@ -27810,66 +31884,115 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntr getRangesBuilderList() { return getRangesFieldBuilder().getBuilderList(); } - private com.google.protobuf.RepeatedFieldBuilder< + private com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryRange, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryRange.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryRangeOrBuilder> getRangesFieldBuilder() { if (rangesBuilder_ == null) { - rangesBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< + rangesBuilder_ = new com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryRange, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryRange.Builder, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryRangeOrBuilder>( ranges_, - ((bitField0_ & 0x00000008) == 0x00000008), + ((bitField0_ & 0x00000008) != 0), getParentForChildren(), isClean()); ranges_ = null; } return rangesBuilder_; } + @java.lang.Override + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:CacheEntry) } + // @@protoc_insertion_point(class_scope:CacheEntry) + private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntry DEFAULT_INSTANCE; static { - defaultInstance = new CacheEntry(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntry(); + } + + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntry getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser + PARSER = new com.google.protobuf.AbstractParser() { + @java.lang.Override + public CacheEntry parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new CacheEntry(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntry getDefaultInstanceForType() { + return DEFAULT_INSTANCE; } - // @@protoc_insertion_point(class_scope:CacheEntry) } - public interface CacheTagOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface CacheTagOrBuilder extends + // @@protoc_insertion_point(interface_extends:CacheTag) + com.google.protobuf.MessageOrBuilder { - // optional string table_name = 1; /** * optional string table_name = 1; + * @return Whether the tableName field is set. */ boolean hasTableName(); /** * optional string table_name = 1; + * @return The tableName. */ java.lang.String getTableName(); /** * optional string table_name = 1; + * @return The bytes for tableName. */ com.google.protobuf.ByteString getTableNameBytes(); - // repeated string partition_desc = 2; /** * repeated string partition_desc = 2; + * @return A list containing the partitionDesc. */ java.util.List - getPartitionDescList(); + getPartitionDescList(); /** * repeated string partition_desc = 2; + * @return The count of partitionDesc. */ int getPartitionDescCount(); /** * repeated string partition_desc = 2; + * @param index The index of the element to return. + * @return The partitionDesc at the given index. */ java.lang.String getPartitionDesc(int index); /** * repeated string partition_desc = 2; + * @param index The index of the value to return. + * @return The bytes of the partitionDesc at the given index. */ com.google.protobuf.ByteString getPartitionDescBytes(int index); @@ -27878,35 +32001,39 @@ public interface CacheTagOrBuilder * Protobuf type {@code CacheTag} */ public static final class CacheTag extends - com.google.protobuf.GeneratedMessage - implements CacheTagOrBuilder { + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:CacheTag) + CacheTagOrBuilder { + private static final long serialVersionUID = 0L; // Use CacheTag.newBuilder() to construct. - private CacheTag(com.google.protobuf.GeneratedMessage.Builder builder) { + private CacheTag(com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private CacheTag(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final CacheTag defaultInstance; - public static CacheTag getDefaultInstance() { - return defaultInstance; + private CacheTag() { + tableName_ = ""; + partitionDesc_ = com.google.protobuf.LazyStringArrayList.EMPTY; } - public CacheTag getDefaultInstanceForType() { - return defaultInstance; + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance( + UnusedPrivateParameter unused) { + return new CacheTag(); } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private CacheTag( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); @@ -27918,24 +32045,26 @@ private CacheTag( case 0: done = true; break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } case 10: { + com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000001; - tableName_ = input.readBytes(); + tableName_ = bs; break; } case 18: { - if (!((mutable_bitField0_ & 0x00000002) == 0x00000002)) { + com.google.protobuf.ByteString bs = input.readBytes(); + if (!((mutable_bitField0_ & 0x00000002) != 0)) { partitionDesc_ = new com.google.protobuf.LazyStringArrayList(); mutable_bitField0_ |= 0x00000002; } - partitionDesc_.add(input.readBytes()); + partitionDesc_.add(bs); + break; + } + default: { + if (!parseUnknownField( + input, unknownFields, extensionRegistry, tag)) { + done = true; + } break; } } @@ -27944,10 +32073,10 @@ private CacheTag( throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { - if (((mutable_bitField0_ & 0x00000002) == 0x00000002)) { - partitionDesc_ = new com.google.protobuf.UnmodifiableLazyStringList(partitionDesc_); + if (((mutable_bitField0_ & 0x00000002) != 0)) { + partitionDesc_ = partitionDesc_.getUnmodifiableView(); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); @@ -27958,41 +32087,30 @@ private CacheTag( return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_CacheTag_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_CacheTag_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public CacheTag parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new CacheTag(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - private int bitField0_; - // optional string table_name = 1; public static final int TABLE_NAME_FIELD_NUMBER = 1; - private java.lang.Object tableName_; + private volatile java.lang.Object tableName_; /** * optional string table_name = 1; + * @return Whether the tableName field is set. */ + @java.lang.Override public boolean hasTableName() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional string table_name = 1; + * @return The tableName. */ + @java.lang.Override public java.lang.String getTableName() { java.lang.Object ref = tableName_; if (ref instanceof java.lang.String) { @@ -28009,7 +32127,9 @@ public java.lang.String getTableName() { } /** * optional string table_name = 1; + * @return The bytes for tableName. */ + @java.lang.Override public com.google.protobuf.ByteString getTableNameBytes() { java.lang.Object ref = tableName_; @@ -28024,92 +32144,86 @@ public java.lang.String getTableName() { } } - // repeated string partition_desc = 2; public static final int PARTITION_DESC_FIELD_NUMBER = 2; private com.google.protobuf.LazyStringList partitionDesc_; /** * repeated string partition_desc = 2; + * @return A list containing the partitionDesc. */ - public java.util.List + public com.google.protobuf.ProtocolStringList getPartitionDescList() { return partitionDesc_; } /** * repeated string partition_desc = 2; + * @return The count of partitionDesc. */ public int getPartitionDescCount() { return partitionDesc_.size(); } /** * repeated string partition_desc = 2; + * @param index The index of the element to return. + * @return The partitionDesc at the given index. */ public java.lang.String getPartitionDesc(int index) { return partitionDesc_.get(index); } /** * repeated string partition_desc = 2; + * @param index The index of the value to return. + * @return The bytes of the partitionDesc at the given index. */ public com.google.protobuf.ByteString getPartitionDescBytes(int index) { return partitionDesc_.getByteString(index); } - private void initFields() { - tableName_ = ""; - partitionDesc_ = com.google.protobuf.LazyStringArrayList.EMPTY; - } private byte memoizedIsInitialized = -1; + @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } + @java.lang.Override public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); - if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeBytes(1, getTableNameBytes()); + if (((bitField0_ & 0x00000001) != 0)) { + com.google.protobuf.GeneratedMessageV3.writeString(output, 1, tableName_); } for (int i = 0; i < partitionDesc_.size(); i++) { - output.writeBytes(2, partitionDesc_.getByteString(i)); + com.google.protobuf.GeneratedMessageV3.writeString(output, 2, partitionDesc_.getRaw(i)); } - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; + @java.lang.Override public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(1, getTableNameBytes()); + if (((bitField0_ & 0x00000001) != 0)) { + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, tableName_); } { int dataSize = 0; for (int i = 0; i < partitionDesc_.size(); i++) { - dataSize += com.google.protobuf.CodedOutputStream - .computeBytesSizeNoTag(partitionDesc_.getByteString(i)); + dataSize += computeStringSizeNoTag(partitionDesc_.getRaw(i)); } size += dataSize; size += 1 * getPartitionDescList().size(); } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { @@ -28120,27 +32234,24 @@ public boolean equals(final java.lang.Object obj) { } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag) obj; - boolean result = true; - result = result && (hasTableName() == other.hasTableName()); + if (hasTableName() != other.hasTableName()) return false; if (hasTableName()) { - result = result && getTableName() - .equals(other.getTableName()); + if (!getTableName() + .equals(other.getTableName())) return false; } - result = result && getPartitionDescList() - .equals(other.getPartitionDescList()); - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; + if (!getPartitionDescList() + .equals(other.getPartitionDescList())) return false; + if (!unknownFields.equals(other.unknownFields)) return false; + return true; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); + hash = (19 * hash) + getDescriptor().hashCode(); if (hasTableName()) { hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER; hash = (53 * hash) + getTableName().hashCode(); @@ -28149,11 +32260,22 @@ public int hashCode() { hash = (37 * hash) + PARTITION_DESC_FIELD_NUMBER; hash = (53 * hash) + getPartitionDescList().hashCode(); } - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag parseFrom( + java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag parseFrom( + java.nio.ByteBuffer data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { @@ -28177,46 +32299,59 @@ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.Ca } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } + @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -28224,14 +32359,16 @@ protected Builder newBuilderForType( * Protobuf type {@code CacheTag} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTagOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder implements + // @@protoc_insertion_point(builder_implements:CacheTag) + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTagOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_CacheTag_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_CacheTag_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -28244,18 +32381,16 @@ private Builder() { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { } } - private static Builder create() { - return new Builder(); - } - + @java.lang.Override public Builder clear() { super.clear(); tableName_ = ""; @@ -28265,19 +32400,18 @@ public Builder clear() { return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - + @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_CacheTag_descriptor; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag.getDefaultInstance(); } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag result = buildPartial(); if (!result.isInitialized()) { @@ -28286,17 +32420,17 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag return result; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { + if (((from_bitField0_ & 0x00000001) != 0)) { to_bitField0_ |= 0x00000001; } result.tableName_ = tableName_; - if (((bitField0_ & 0x00000002) == 0x00000002)) { - partitionDesc_ = new com.google.protobuf.UnmodifiableLazyStringList( - partitionDesc_); + if (((bitField0_ & 0x00000002) != 0)) { + partitionDesc_ = partitionDesc_.getUnmodifiableView(); bitField0_ = (bitField0_ & ~0x00000002); } result.partitionDesc_ = partitionDesc_; @@ -28305,6 +32439,39 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag return result; } + @java.lang.Override + public Builder clone() { + return super.clone(); + } + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.setField(field, value); + } + @java.lang.Override + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + @java.lang.Override + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.addRepeatedField(field, value); + } + @java.lang.Override public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag)other); @@ -28331,14 +32498,17 @@ public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtoc } onChanged(); } - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } + @java.lang.Override public final boolean isInitialized() { return true; } + @java.lang.Override public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) @@ -28348,7 +32518,7 @@ public Builder mergeFrom( parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -28358,23 +32528,27 @@ public Builder mergeFrom( } private int bitField0_; - // optional string table_name = 1; private java.lang.Object tableName_ = ""; /** * optional string table_name = 1; + * @return Whether the tableName field is set. */ public boolean hasTableName() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional string table_name = 1; + * @return The tableName. */ public java.lang.String getTableName() { java.lang.Object ref = tableName_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - tableName_ = s; + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + tableName_ = s; + } return s; } else { return (java.lang.String) ref; @@ -28382,6 +32556,7 @@ public java.lang.String getTableName() { } /** * optional string table_name = 1; + * @return The bytes for tableName. */ public com.google.protobuf.ByteString getTableNameBytes() { @@ -28398,6 +32573,8 @@ public java.lang.String getTableName() { } /** * optional string table_name = 1; + * @param value The tableName to set. + * @return This builder for chaining. */ public Builder setTableName( java.lang.String value) { @@ -28411,6 +32588,7 @@ public Builder setTableName( } /** * optional string table_name = 1; + * @return This builder for chaining. */ public Builder clearTableName() { bitField0_ = (bitField0_ & ~0x00000001); @@ -28420,6 +32598,8 @@ public Builder clearTableName() { } /** * optional string table_name = 1; + * @param value The bytes for tableName to set. + * @return This builder for chaining. */ public Builder setTableNameBytes( com.google.protobuf.ByteString value) { @@ -28432,35 +32612,40 @@ public Builder setTableNameBytes( return this; } - // repeated string partition_desc = 2; private com.google.protobuf.LazyStringList partitionDesc_ = com.google.protobuf.LazyStringArrayList.EMPTY; private void ensurePartitionDescIsMutable() { - if (!((bitField0_ & 0x00000002) == 0x00000002)) { + if (!((bitField0_ & 0x00000002) != 0)) { partitionDesc_ = new com.google.protobuf.LazyStringArrayList(partitionDesc_); bitField0_ |= 0x00000002; } } /** * repeated string partition_desc = 2; + * @return A list containing the partitionDesc. */ - public java.util.List + public com.google.protobuf.ProtocolStringList getPartitionDescList() { - return java.util.Collections.unmodifiableList(partitionDesc_); + return partitionDesc_.getUnmodifiableView(); } /** * repeated string partition_desc = 2; + * @return The count of partitionDesc. */ public int getPartitionDescCount() { return partitionDesc_.size(); } /** * repeated string partition_desc = 2; + * @param index The index of the element to return. + * @return The partitionDesc at the given index. */ public java.lang.String getPartitionDesc(int index) { return partitionDesc_.get(index); } /** * repeated string partition_desc = 2; + * @param index The index of the value to return. + * @return The bytes of the partitionDesc at the given index. */ public com.google.protobuf.ByteString getPartitionDescBytes(int index) { @@ -28468,6 +32653,9 @@ public java.lang.String getPartitionDesc(int index) { } /** * repeated string partition_desc = 2; + * @param index The index to set the value at. + * @param value The partitionDesc to set. + * @return This builder for chaining. */ public Builder setPartitionDesc( int index, java.lang.String value) { @@ -28481,6 +32669,8 @@ public Builder setPartitionDesc( } /** * repeated string partition_desc = 2; + * @param value The partitionDesc to add. + * @return This builder for chaining. */ public Builder addPartitionDesc( java.lang.String value) { @@ -28494,16 +32684,20 @@ public Builder addPartitionDesc( } /** * repeated string partition_desc = 2; + * @param values The partitionDesc to add. + * @return This builder for chaining. */ public Builder addAllPartitionDesc( java.lang.Iterable values) { ensurePartitionDescIsMutable(); - super.addAll(values, partitionDesc_); + com.google.protobuf.AbstractMessageLite.Builder.addAll( + values, partitionDesc_); onChanged(); return this; } /** * repeated string partition_desc = 2; + * @return This builder for chaining. */ public Builder clearPartitionDesc() { partitionDesc_ = com.google.protobuf.LazyStringArrayList.EMPTY; @@ -28513,6 +32707,8 @@ public Builder clearPartitionDesc() { } /** * repeated string partition_desc = 2; + * @param value The bytes of the partitionDesc to add. + * @return This builder for chaining. */ public Builder addPartitionDescBytes( com.google.protobuf.ByteString value) { @@ -28524,38 +32720,82 @@ public Builder addPartitionDescBytes( onChanged(); return this; } + @java.lang.Override + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:CacheTag) } + // @@protoc_insertion_point(class_scope:CacheTag) + private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag DEFAULT_INSTANCE; static { - defaultInstance = new CacheTag(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag(); + } + + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser + PARSER = new com.google.protobuf.AbstractParser() { + @java.lang.Override + public CacheTag parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new CacheTag(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheTag getDefaultInstanceForType() { + return DEFAULT_INSTANCE; } - // @@protoc_insertion_point(class_scope:CacheTag) } - public interface CacheEntryRangeOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface CacheEntryRangeOrBuilder extends + // @@protoc_insertion_point(interface_extends:CacheEntryRange) + com.google.protobuf.MessageOrBuilder { - // optional int64 start = 1; /** * optional int64 start = 1; + * @return Whether the start field is set. */ boolean hasStart(); /** * optional int64 start = 1; + * @return The start. */ long getStart(); - // optional int64 end = 2; /** * optional int64 end = 2; + * @return Whether the end field is set. */ boolean hasEnd(); /** * optional int64 end = 2; + * @return The end. */ long getEnd(); } @@ -28563,35 +32803,37 @@ public interface CacheEntryRangeOrBuilder * Protobuf type {@code CacheEntryRange} */ public static final class CacheEntryRange extends - com.google.protobuf.GeneratedMessage - implements CacheEntryRangeOrBuilder { + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:CacheEntryRange) + CacheEntryRangeOrBuilder { + private static final long serialVersionUID = 0L; // Use CacheEntryRange.newBuilder() to construct. - private CacheEntryRange(com.google.protobuf.GeneratedMessage.Builder builder) { + private CacheEntryRange(com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private CacheEntryRange(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final CacheEntryRange defaultInstance; - public static CacheEntryRange getDefaultInstance() { - return defaultInstance; + private CacheEntryRange() { } - public CacheEntryRange getDefaultInstanceForType() { - return defaultInstance; + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance( + UnusedPrivateParameter unused) { + return new CacheEntryRange(); } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private CacheEntryRange( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); @@ -28603,13 +32845,6 @@ private CacheEntryRange( case 0: done = true; break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } case 8: { bitField0_ |= 0x00000001; start_ = input.readInt64(); @@ -28620,13 +32855,20 @@ private CacheEntryRange( end_ = input.readInt64(); break; } + default: { + if (!parseUnknownField( + input, unknownFields, extensionRegistry, tag)) { + done = true; + } + break; + } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); @@ -28637,112 +32879,95 @@ private CacheEntryRange( return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_CacheEntryRange_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_CacheEntryRange_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryRange.class, org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryRange.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public CacheEntryRange parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new CacheEntryRange(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - private int bitField0_; - // optional int64 start = 1; public static final int START_FIELD_NUMBER = 1; private long start_; /** * optional int64 start = 1; + * @return Whether the start field is set. */ + @java.lang.Override public boolean hasStart() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional int64 start = 1; + * @return The start. */ + @java.lang.Override public long getStart() { return start_; } - // optional int64 end = 2; public static final int END_FIELD_NUMBER = 2; private long end_; /** * optional int64 end = 2; + * @return Whether the end field is set. */ + @java.lang.Override public boolean hasEnd() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * optional int64 end = 2; + * @return The end. */ + @java.lang.Override public long getEnd() { return end_; } - private void initFields() { - start_ = 0L; - end_ = 0L; - } private byte memoizedIsInitialized = -1; + @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } + @java.lang.Override public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); - if (((bitField0_ & 0x00000001) == 0x00000001)) { + if (((bitField0_ & 0x00000001) != 0)) { output.writeInt64(1, start_); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { + if (((bitField0_ & 0x00000002) != 0)) { output.writeInt64(2, end_); } - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; + @java.lang.Override public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { + if (((bitField0_ & 0x00000001) != 0)) { size += com.google.protobuf.CodedOutputStream .computeInt64Size(1, start_); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { + if (((bitField0_ & 0x00000002) != 0)) { size += com.google.protobuf.CodedOutputStream .computeInt64Size(2, end_); } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { @@ -28753,43 +32978,53 @@ public boolean equals(final java.lang.Object obj) { } org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryRange other = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryRange) obj; - boolean result = true; - result = result && (hasStart() == other.hasStart()); + if (hasStart() != other.hasStart()) return false; if (hasStart()) { - result = result && (getStart() - == other.getStart()); + if (getStart() + != other.getStart()) return false; } - result = result && (hasEnd() == other.hasEnd()); + if (hasEnd() != other.hasEnd()) return false; if (hasEnd()) { - result = result && (getEnd() - == other.getEnd()); + if (getEnd() + != other.getEnd()) return false; } - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; + if (!unknownFields.equals(other.unknownFields)) return false; + return true; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); + hash = (19 * hash) + getDescriptor().hashCode(); if (hasStart()) { hash = (37 * hash) + START_FIELD_NUMBER; - hash = (53 * hash) + hashLong(getStart()); + hash = (53 * hash) + com.google.protobuf.Internal.hashLong( + getStart()); } if (hasEnd()) { hash = (37 * hash) + END_FIELD_NUMBER; - hash = (53 * hash) + hashLong(getEnd()); + hash = (53 * hash) + com.google.protobuf.Internal.hashLong( + getEnd()); } - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryRange parseFrom( + java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryRange parseFrom( + java.nio.ByteBuffer data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryRange parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { @@ -28813,46 +33048,59 @@ public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.Ca } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryRange parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryRange parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryRange parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryRange parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryRange parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryRange parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } + @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryRange prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -28860,14 +33108,16 @@ protected Builder newBuilderForType( * Protobuf type {@code CacheEntryRange} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryRangeOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder implements + // @@protoc_insertion_point(builder_implements:CacheEntryRange) + org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryRangeOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_CacheEntryRange_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_CacheEntryRange_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -28880,18 +33130,16 @@ private Builder() { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { } } - private static Builder create() { - return new Builder(); - } - + @java.lang.Override public Builder clear() { super.clear(); start_ = 0L; @@ -28901,19 +33149,18 @@ public Builder clear() { return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - + @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.internal_static_CacheEntryRange_descriptor; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryRange getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryRange.getDefaultInstance(); } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryRange build() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryRange result = buildPartial(); if (!result.isInitialized()) { @@ -28922,23 +33169,57 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntr return result; } + @java.lang.Override public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryRange buildPartial() { org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryRange result = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryRange(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { + if (((from_bitField0_ & 0x00000001) != 0)) { + result.start_ = start_; to_bitField0_ |= 0x00000001; } - result.start_ = start_; - if (((from_bitField0_ & 0x00000002) == 0x00000002)) { + if (((from_bitField0_ & 0x00000002) != 0)) { + result.end_ = end_; to_bitField0_ |= 0x00000002; } - result.end_ = end_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } + @java.lang.Override + public Builder clone() { + return super.clone(); + } + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.setField(field, value); + } + @java.lang.Override + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + @java.lang.Override + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.addRepeatedField(field, value); + } + @java.lang.Override public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryRange) { return mergeFrom((org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryRange)other); @@ -28956,14 +33237,17 @@ public Builder mergeFrom(org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtoc if (other.hasEnd()) { setEnd(other.getEnd()); } - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } + @java.lang.Override public final boolean isInitialized() { return true; } + @java.lang.Override public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) @@ -28973,7 +33257,7 @@ public Builder mergeFrom( parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryRange) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -28983,22 +33267,27 @@ public Builder mergeFrom( } private int bitField0_; - // optional int64 start = 1; private long start_ ; /** * optional int64 start = 1; + * @return Whether the start field is set. */ + @java.lang.Override public boolean hasStart() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional int64 start = 1; + * @return The start. */ + @java.lang.Override public long getStart() { return start_; } /** * optional int64 start = 1; + * @param value The start to set. + * @return This builder for chaining. */ public Builder setStart(long value) { bitField0_ |= 0x00000001; @@ -29008,6 +33297,7 @@ public Builder setStart(long value) { } /** * optional int64 start = 1; + * @return This builder for chaining. */ public Builder clearStart() { bitField0_ = (bitField0_ & ~0x00000001); @@ -29016,22 +33306,27 @@ public Builder clearStart() { return this; } - // optional int64 end = 2; private long end_ ; /** * optional int64 end = 2; + * @return Whether the end field is set. */ + @java.lang.Override public boolean hasEnd() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * optional int64 end = 2; + * @return The end. */ + @java.lang.Override public long getEnd() { return end_; } /** * optional int64 end = 2; + * @param value The end to set. + * @return This builder for chaining. */ public Builder setEnd(long value) { bitField0_ |= 0x00000002; @@ -29041,6 +33336,7 @@ public Builder setEnd(long value) { } /** * optional int64 end = 2; + * @return This builder for chaining. */ public Builder clearEnd() { bitField0_ = (bitField0_ & ~0x00000002); @@ -29048,16 +33344,57 @@ public Builder clearEnd() { onChanged(); return this; } + @java.lang.Override + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:CacheEntryRange) } + // @@protoc_insertion_point(class_scope:CacheEntryRange) + private static final org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryRange DEFAULT_INSTANCE; static { - defaultInstance = new CacheEntryRange(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryRange(); + } + + public static org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryRange getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser + PARSER = new com.google.protobuf.AbstractParser() { + @java.lang.Override + public CacheEntryRange parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new CacheEntryRange(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.CacheEntryRange getDefaultInstanceForType() { + return DEFAULT_INSTANCE; } - // @@protoc_insertion_point(class_scope:CacheEntryRange) } /** @@ -30234,212 +34571,212 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheC // @@protoc_insertion_point(class_scope:LlapManagementProtocol) } - private static com.google.protobuf.Descriptors.Descriptor + private static final com.google.protobuf.Descriptors.Descriptor internal_static_UserPayloadProto_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + private static final + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_UserPayloadProto_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static final com.google.protobuf.Descriptors.Descriptor internal_static_EntityDescriptorProto_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + private static final + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_EntityDescriptorProto_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static final com.google.protobuf.Descriptors.Descriptor internal_static_IOSpecProto_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + private static final + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_IOSpecProto_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static final com.google.protobuf.Descriptors.Descriptor internal_static_GroupInputSpecProto_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + private static final + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_GroupInputSpecProto_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static final com.google.protobuf.Descriptors.Descriptor internal_static_SignableVertexSpec_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + private static final + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_SignableVertexSpec_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static final com.google.protobuf.Descriptors.Descriptor internal_static_VertexOrBinary_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + private static final + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_VertexOrBinary_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static final com.google.protobuf.Descriptors.Descriptor internal_static_FragmentRuntimeInfo_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + private static final + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_FragmentRuntimeInfo_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static final com.google.protobuf.Descriptors.Descriptor internal_static_QueryIdentifierProto_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + private static final + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_QueryIdentifierProto_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static final com.google.protobuf.Descriptors.Descriptor internal_static_NotTezEvent_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + private static final + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_NotTezEvent_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static final com.google.protobuf.Descriptors.Descriptor internal_static_SubmitWorkRequestProto_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + private static final + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_SubmitWorkRequestProto_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static final com.google.protobuf.Descriptors.Descriptor internal_static_RegisterDagRequestProto_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + private static final + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_RegisterDagRequestProto_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static final com.google.protobuf.Descriptors.Descriptor internal_static_RegisterDagResponseProto_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + private static final + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_RegisterDagResponseProto_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static final com.google.protobuf.Descriptors.Descriptor internal_static_SubmitWorkResponseProto_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + private static final + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_SubmitWorkResponseProto_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static final com.google.protobuf.Descriptors.Descriptor internal_static_SourceStateUpdatedRequestProto_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + private static final + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_SourceStateUpdatedRequestProto_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static final com.google.protobuf.Descriptors.Descriptor internal_static_SourceStateUpdatedResponseProto_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + private static final + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_SourceStateUpdatedResponseProto_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static final com.google.protobuf.Descriptors.Descriptor internal_static_QueryCompleteRequestProto_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + private static final + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_QueryCompleteRequestProto_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static final com.google.protobuf.Descriptors.Descriptor internal_static_QueryCompleteResponseProto_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + private static final + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_QueryCompleteResponseProto_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static final com.google.protobuf.Descriptors.Descriptor internal_static_TerminateFragmentRequestProto_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + private static final + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_TerminateFragmentRequestProto_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static final com.google.protobuf.Descriptors.Descriptor internal_static_TerminateFragmentResponseProto_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + private static final + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_TerminateFragmentResponseProto_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static final com.google.protobuf.Descriptors.Descriptor internal_static_UpdateFragmentRequestProto_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + private static final + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_UpdateFragmentRequestProto_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static final com.google.protobuf.Descriptors.Descriptor internal_static_UpdateFragmentResponseProto_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + private static final + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_UpdateFragmentResponseProto_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static final com.google.protobuf.Descriptors.Descriptor internal_static_GetTokenRequestProto_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + private static final + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_GetTokenRequestProto_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static final com.google.protobuf.Descriptors.Descriptor internal_static_GetTokenResponseProto_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + private static final + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_GetTokenResponseProto_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static final com.google.protobuf.Descriptors.Descriptor internal_static_LlapOutputSocketInitMessage_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + private static final + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_LlapOutputSocketInitMessage_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static final com.google.protobuf.Descriptors.Descriptor internal_static_PurgeCacheRequestProto_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + private static final + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_PurgeCacheRequestProto_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static final com.google.protobuf.Descriptors.Descriptor internal_static_PurgeCacheResponseProto_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + private static final + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_PurgeCacheResponseProto_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static final com.google.protobuf.Descriptors.Descriptor internal_static_MapEntry_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + private static final + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_MapEntry_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static final com.google.protobuf.Descriptors.Descriptor internal_static_GetDaemonMetricsRequestProto_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + private static final + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_GetDaemonMetricsRequestProto_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static final com.google.protobuf.Descriptors.Descriptor internal_static_GetDaemonMetricsResponseProto_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + private static final + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_GetDaemonMetricsResponseProto_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static final com.google.protobuf.Descriptors.Descriptor internal_static_SetCapacityRequestProto_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + private static final + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_SetCapacityRequestProto_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static final com.google.protobuf.Descriptors.Descriptor internal_static_SetCapacityResponseProto_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + private static final + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_SetCapacityResponseProto_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static final com.google.protobuf.Descriptors.Descriptor internal_static_EvictEntityRequestProto_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + private static final + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_EvictEntityRequestProto_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static final com.google.protobuf.Descriptors.Descriptor internal_static_TableProto_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + private static final + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_TableProto_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static final com.google.protobuf.Descriptors.Descriptor internal_static_EvictEntityResponseProto_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + private static final + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_EvictEntityResponseProto_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static final com.google.protobuf.Descriptors.Descriptor internal_static_GetCacheContentRequestProto_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + private static final + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_GetCacheContentRequestProto_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static final com.google.protobuf.Descriptors.Descriptor internal_static_GetCacheContentResponseProto_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + private static final + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_GetCacheContentResponseProto_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static final com.google.protobuf.Descriptors.Descriptor internal_static_CacheEntryList_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + private static final + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_CacheEntryList_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static final com.google.protobuf.Descriptors.Descriptor internal_static_CacheEntry_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + private static final + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_CacheEntry_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static final com.google.protobuf.Descriptors.Descriptor internal_static_CacheTag_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + private static final + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_CacheTag_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static final com.google.protobuf.Descriptors.Descriptor internal_static_CacheEntryRange_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + private static final + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_CacheEntryRange_fieldAccessorTable; public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { return descriptor; } - private static com.google.protobuf.Descriptors.FileDescriptor + private static com.google.protobuf.Descriptors.FileDescriptor descriptor; static { java.lang.String[] descriptorData = { @@ -30452,7 +34789,7 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheC "-\n\rio_descriptor\030\002 \001(\0132\026.EntityDescripto" + "rProto\022\033\n\023physical_edge_count\030\003 \001(\005\"z\n\023G" + "roupInputSpecProto\022\022\n\ngroup_name\030\001 \001(\t\022\026" + - "\n\016group_vertices\030\002 \003(\t\0227\n\027merged_input_d", + "\n\016group_vertices\030\002 \003(\t\0227\n\027merged_input_d" + "escriptor\030\003 \001(\0132\026.EntityDescriptorProto\"" + "\314\003\n\022SignableVertexSpec\022\014\n\004user\030\001 \001(\t\022\026\n\016" + "signatureKeyId\030\002 \001(\003\022/\n\020query_identifier" + @@ -30462,7 +34799,7 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheC "en_identifier\030\010 \001(\t\0224\n\024processor_descrip" + "tor\030\t \001(\0132\026.EntityDescriptorProto\022!\n\013inp" + "ut_specs\030\n \003(\0132\014.IOSpecProto\022\"\n\014output_s" + - "pecs\030\013 \003(\0132\014.IOSpecProto\0221\n\023grouped_inpu", + "pecs\030\013 \003(\0132\014.IOSpecProto\0221\n\023grouped_inpu" + "t_specs\030\014 \003(\0132\024.GroupInputSpecProto\022\032\n\022v" + "ertex_parallelism\030\r \001(\005\022%\n\026is_external_s" + "ubmission\030\016 \001(\010:\005false\"K\n\016VertexOrBinary" + @@ -30472,7 +34809,7 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheC "\022-\n%num_self_and_upstream_completed_task" + "s\030\002 \001(\005\022\033\n\023within_dag_priority\030\003 \001(\005\022\026\n\016" + "dag_start_time\030\004 \001(\003\022 \n\030first_attempt_st" + - "art_time\030\005 \001(\003\022\"\n\032current_attempt_start_", + "art_time\030\005 \001(\003\022\"\n\032current_attempt_start_" + "time\030\006 \001(\003\"d\n\024QueryIdentifierProto\022\035\n\025ap" + "plication_id_string\030\001 \001(\t\022\021\n\tdag_index\030\002" + " \001(\005\022\032\n\022app_attempt_number\030\003 \001(\005\"l\n\013NotT" + @@ -30482,7 +34819,7 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheC "uestProto\022\"\n\twork_spec\030\001 \001(\0132\017.VertexOrB" + "inary\022\033\n\023work_spec_signature\030\002 \001(\014\022\027\n\017fr" + "agment_number\030\003 \001(\005\022\026\n\016attempt_number\030\004 " + - "\001(\005\022\033\n\023container_id_string\030\005 \001(\t\022\017\n\007am_h", + "\001(\005\022\033\n\023container_id_string\030\005 \001(\t\022\017\n\007am_h" + "ost\030\006 \001(\t\022\017\n\007am_port\030\007 \001(\005\022\032\n\022credential" + "s_binary\030\010 \001(\014\0223\n\025fragment_runtime_info\030" + "\t \001(\0132\024.FragmentRuntimeInfo\022\033\n\023initial_e" + @@ -30492,7 +34829,7 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheC "st\030\016 \001(\010:\005false\"t\n\027RegisterDagRequestPro" + "to\022\014\n\004user\030\001 \001(\t\022/\n\020query_identifier\030\002 \002" + "(\0132\025.QueryIdentifierProto\022\032\n\022credentials" + - "_binary\030\003 \001(\014\"\032\n\030RegisterDagResponseProt", + "_binary\030\003 \001(\014\"\032\n\030RegisterDagResponseProt" + "o\"b\n\027SubmitWorkResponseProto\022/\n\020submissi" + "on_state\030\001 \001(\0162\025.SubmissionStateProto\022\026\n" + "\016unique_node_id\030\002 \001(\t\"\205\001\n\036SourceStateUpd" + @@ -30502,7 +34839,7 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheC "\037SourceStateUpdatedResponseProto\"e\n\031Quer" + "yCompleteRequestProto\022/\n\020query_identifie" + "r\030\001 \001(\0132\025.QueryIdentifierProto\022\027\n\014delete" + - "_delay\030\002 \001(\003:\0010\"\034\n\032QueryCompleteResponse", + "_delay\030\002 \001(\003:\0010\"\034\n\032QueryCompleteResponse" + "Proto\"t\n\035TerminateFragmentRequestProto\022/" + "\n\020query_identifier\030\001 \001(\0132\025.QueryIdentifi" + "erProto\022\"\n\032fragment_identifier_string\030\002 " + @@ -30512,7 +34849,7 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheC "\032fragment_identifier_string\030\002 \001(\t\022\025\n\ris_" + "guaranteed\030\003 \001(\010\"D\n\033UpdateFragmentRespon" + "seProto\022\016\n\006result\030\001 \001(\010\022\025\n\ris_guaranteed" + - "\030\002 \001(\010\"&\n\024GetTokenRequestProto\022\016\n\006app_id", + "\030\002 \001(\010\"&\n\024GetTokenRequestProto\022\016\n\006app_id" + "\030\001 \001(\t\"&\n\025GetTokenResponseProto\022\r\n\005token" + "\030\001 \001(\014\"A\n\033LlapOutputSocketInitMessage\022\023\n" + "\013fragment_id\030\001 \002(\t\022\r\n\005token\030\002 \001(\014\"\030\n\026Pur" + @@ -30522,7 +34859,7 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheC "etDaemonMetricsRequestProto\";\n\035GetDaemon" + "MetricsResponseProto\022\032\n\007metrics\030\001 \003(\0132\t." + "MapEntry\"A\n\027SetCapacityRequestProto\022\023\n\013e" + - "xecutorNum\030\001 \001(\005\022\021\n\tqueueSize\030\002 \001(\005\"\032\n\030S", + "xecutorNum\030\001 \001(\005\022\021\n\tqueueSize\030\002 \001(\005\"\032\n\030S" + "etCapacityResponseProto\"F\n\027EvictEntityRe" + "questProto\022\017\n\007db_name\030\001 \002(\t\022\032\n\005table\030\002 \003" + "(\0132\013.TableProto\"D\n\nTableProto\022\022\n\ntable_n" + @@ -30532,7 +34869,7 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheC "stProto\"?\n\034GetCacheContentResponseProto\022" + "\037\n\006result\030\001 \001(\0132\017.CacheEntryList\".\n\016Cach" + "eEntryList\022\034\n\007entries\030\001 \003(\0132\013.CacheEntry" + - "\"q\n\nCacheEntry\022\020\n\010file_key\030\001 \001(\014\022\021\n\tfile", + "\"q\n\nCacheEntry\022\020\n\010file_key\030\001 \001(\014\022\021\n\tfile" + "_path\030\002 \001(\t\022\034\n\tcache_tag\030\003 \001(\0132\t.CacheTa" + "g\022 \n\006ranges\030\004 \003(\0132\020.CacheEntryRange\"6\n\010C" + "acheTag\022\022\n\ntable_name\030\001 \001(\t\022\026\n\016partition" + @@ -30542,7 +34879,7 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheC "sionStateProto\022\014\n\010ACCEPTED\020\001\022\014\n\010REJECTED" + "\020\002\022\021\n\rEVICTED_OTHER\020\0032\337\003\n\022LlapDaemonProt" + "ocol\022B\n\013registerDag\022\030.RegisterDagRequest" + - "Proto\032\031.RegisterDagResponseProto\022?\n\nsubm", + "Proto\032\031.RegisterDagResponseProto\022?\n\nsubm" + "itWork\022\027.SubmitWorkRequestProto\032\030.Submit" + "WorkResponseProto\022W\n\022sourceStateUpdated\022" + "\037.SourceStateUpdatedRequestProto\032 .Sourc" + @@ -30552,7 +34889,7 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheC "t\022\036.TerminateFragmentRequestProto\032\037.Term" + "inateFragmentResponseProto\022K\n\016updateFrag" + "ment\022\033.UpdateFragmentRequestProto\032\034.Upda" + - "teFragmentResponseProto2\311\003\n\026LlapManageme", + "teFragmentResponseProto2\311\003\n\026LlapManageme" + "ntProtocol\022C\n\022getDelegationToken\022\025.GetTo" + "kenRequestProto\032\026.GetTokenResponseProto\022" + "?\n\npurgeCache\022\027.PurgeCacheRequestProto\032\030" + @@ -30562,263 +34899,255 @@ public org.apache.hadoop.hive.llap.daemon.rpc.LlapDaemonProtocolProtos.GetCacheC "ty\022\030.SetCapacityRequestProto\032\031.SetCapaci" + "tyResponseProto\022B\n\013evictEntity\022\030.EvictEn" + "tityRequestProto\032\031.EvictEntityResponsePr" + - "oto\022N\n\017getCacheContent\022\034.GetCacheContent", + "oto\022N\n\017getCacheContent\022\034.GetCacheContent" + "RequestProto\032\035.GetCacheContentResponsePr" + "otoBH\n&org.apache.hadoop.hive.llap.daemo" + "n.rpcB\030LlapDaemonProtocolProtos\210\001\001\240\001\001" }; - com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = - new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { - public com.google.protobuf.ExtensionRegistry assignDescriptors( - com.google.protobuf.Descriptors.FileDescriptor root) { - descriptor = root; - internal_static_UserPayloadProto_descriptor = - getDescriptor().getMessageTypes().get(0); - internal_static_UserPayloadProto_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_UserPayloadProto_descriptor, - new java.lang.String[] { "UserPayload", "Version", }); - internal_static_EntityDescriptorProto_descriptor = - getDescriptor().getMessageTypes().get(1); - internal_static_EntityDescriptorProto_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_EntityDescriptorProto_descriptor, - new java.lang.String[] { "ClassName", "UserPayload", "HistoryText", }); - internal_static_IOSpecProto_descriptor = - getDescriptor().getMessageTypes().get(2); - internal_static_IOSpecProto_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_IOSpecProto_descriptor, - new java.lang.String[] { "ConnectedVertexName", "IoDescriptor", "PhysicalEdgeCount", }); - internal_static_GroupInputSpecProto_descriptor = - getDescriptor().getMessageTypes().get(3); - internal_static_GroupInputSpecProto_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_GroupInputSpecProto_descriptor, - new java.lang.String[] { "GroupName", "GroupVertices", "MergedInputDescriptor", }); - internal_static_SignableVertexSpec_descriptor = - getDescriptor().getMessageTypes().get(4); - internal_static_SignableVertexSpec_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_SignableVertexSpec_descriptor, - new java.lang.String[] { "User", "SignatureKeyId", "QueryIdentifier", "HiveQueryId", "DagName", "VertexName", "VertexIndex", "TokenIdentifier", "ProcessorDescriptor", "InputSpecs", "OutputSpecs", "GroupedInputSpecs", "VertexParallelism", "IsExternalSubmission", }); - internal_static_VertexOrBinary_descriptor = - getDescriptor().getMessageTypes().get(5); - internal_static_VertexOrBinary_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_VertexOrBinary_descriptor, - new java.lang.String[] { "Vertex", "VertexBinary", }); - internal_static_FragmentRuntimeInfo_descriptor = - getDescriptor().getMessageTypes().get(6); - internal_static_FragmentRuntimeInfo_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_FragmentRuntimeInfo_descriptor, - new java.lang.String[] { "NumSelfAndUpstreamTasks", "NumSelfAndUpstreamCompletedTasks", "WithinDagPriority", "DagStartTime", "FirstAttemptStartTime", "CurrentAttemptStartTime", }); - internal_static_QueryIdentifierProto_descriptor = - getDescriptor().getMessageTypes().get(7); - internal_static_QueryIdentifierProto_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_QueryIdentifierProto_descriptor, - new java.lang.String[] { "ApplicationIdString", "DagIndex", "AppAttemptNumber", }); - internal_static_NotTezEvent_descriptor = - getDescriptor().getMessageTypes().get(8); - internal_static_NotTezEvent_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_NotTezEvent_descriptor, - new java.lang.String[] { "InputEventProtoBytes", "VertexName", "DestInputName", "KeyId", }); - internal_static_SubmitWorkRequestProto_descriptor = - getDescriptor().getMessageTypes().get(9); - internal_static_SubmitWorkRequestProto_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_SubmitWorkRequestProto_descriptor, - new java.lang.String[] { "WorkSpec", "WorkSpecSignature", "FragmentNumber", "AttemptNumber", "ContainerIdString", "AmHost", "AmPort", "CredentialsBinary", "FragmentRuntimeInfo", "InitialEventBytes", "InitialEventSignature", "IsGuaranteed", "Jwt", "IsExternalClientRequest", }); - internal_static_RegisterDagRequestProto_descriptor = - getDescriptor().getMessageTypes().get(10); - internal_static_RegisterDagRequestProto_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_RegisterDagRequestProto_descriptor, - new java.lang.String[] { "User", "QueryIdentifier", "CredentialsBinary", }); - internal_static_RegisterDagResponseProto_descriptor = - getDescriptor().getMessageTypes().get(11); - internal_static_RegisterDagResponseProto_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_RegisterDagResponseProto_descriptor, - new java.lang.String[] { }); - internal_static_SubmitWorkResponseProto_descriptor = - getDescriptor().getMessageTypes().get(12); - internal_static_SubmitWorkResponseProto_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_SubmitWorkResponseProto_descriptor, - new java.lang.String[] { "SubmissionState", "UniqueNodeId", }); - internal_static_SourceStateUpdatedRequestProto_descriptor = - getDescriptor().getMessageTypes().get(13); - internal_static_SourceStateUpdatedRequestProto_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_SourceStateUpdatedRequestProto_descriptor, - new java.lang.String[] { "QueryIdentifier", "SrcName", "State", }); - internal_static_SourceStateUpdatedResponseProto_descriptor = - getDescriptor().getMessageTypes().get(14); - internal_static_SourceStateUpdatedResponseProto_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_SourceStateUpdatedResponseProto_descriptor, - new java.lang.String[] { }); - internal_static_QueryCompleteRequestProto_descriptor = - getDescriptor().getMessageTypes().get(15); - internal_static_QueryCompleteRequestProto_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_QueryCompleteRequestProto_descriptor, - new java.lang.String[] { "QueryIdentifier", "DeleteDelay", }); - internal_static_QueryCompleteResponseProto_descriptor = - getDescriptor().getMessageTypes().get(16); - internal_static_QueryCompleteResponseProto_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_QueryCompleteResponseProto_descriptor, - new java.lang.String[] { }); - internal_static_TerminateFragmentRequestProto_descriptor = - getDescriptor().getMessageTypes().get(17); - internal_static_TerminateFragmentRequestProto_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_TerminateFragmentRequestProto_descriptor, - new java.lang.String[] { "QueryIdentifier", "FragmentIdentifierString", }); - internal_static_TerminateFragmentResponseProto_descriptor = - getDescriptor().getMessageTypes().get(18); - internal_static_TerminateFragmentResponseProto_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_TerminateFragmentResponseProto_descriptor, - new java.lang.String[] { }); - internal_static_UpdateFragmentRequestProto_descriptor = - getDescriptor().getMessageTypes().get(19); - internal_static_UpdateFragmentRequestProto_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_UpdateFragmentRequestProto_descriptor, - new java.lang.String[] { "QueryIdentifier", "FragmentIdentifierString", "IsGuaranteed", }); - internal_static_UpdateFragmentResponseProto_descriptor = - getDescriptor().getMessageTypes().get(20); - internal_static_UpdateFragmentResponseProto_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_UpdateFragmentResponseProto_descriptor, - new java.lang.String[] { "Result", "IsGuaranteed", }); - internal_static_GetTokenRequestProto_descriptor = - getDescriptor().getMessageTypes().get(21); - internal_static_GetTokenRequestProto_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_GetTokenRequestProto_descriptor, - new java.lang.String[] { "AppId", }); - internal_static_GetTokenResponseProto_descriptor = - getDescriptor().getMessageTypes().get(22); - internal_static_GetTokenResponseProto_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_GetTokenResponseProto_descriptor, - new java.lang.String[] { "Token", }); - internal_static_LlapOutputSocketInitMessage_descriptor = - getDescriptor().getMessageTypes().get(23); - internal_static_LlapOutputSocketInitMessage_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_LlapOutputSocketInitMessage_descriptor, - new java.lang.String[] { "FragmentId", "Token", }); - internal_static_PurgeCacheRequestProto_descriptor = - getDescriptor().getMessageTypes().get(24); - internal_static_PurgeCacheRequestProto_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_PurgeCacheRequestProto_descriptor, - new java.lang.String[] { }); - internal_static_PurgeCacheResponseProto_descriptor = - getDescriptor().getMessageTypes().get(25); - internal_static_PurgeCacheResponseProto_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_PurgeCacheResponseProto_descriptor, - new java.lang.String[] { "PurgedMemoryBytes", }); - internal_static_MapEntry_descriptor = - getDescriptor().getMessageTypes().get(26); - internal_static_MapEntry_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_MapEntry_descriptor, - new java.lang.String[] { "Key", "Value", }); - internal_static_GetDaemonMetricsRequestProto_descriptor = - getDescriptor().getMessageTypes().get(27); - internal_static_GetDaemonMetricsRequestProto_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_GetDaemonMetricsRequestProto_descriptor, - new java.lang.String[] { }); - internal_static_GetDaemonMetricsResponseProto_descriptor = - getDescriptor().getMessageTypes().get(28); - internal_static_GetDaemonMetricsResponseProto_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_GetDaemonMetricsResponseProto_descriptor, - new java.lang.String[] { "Metrics", }); - internal_static_SetCapacityRequestProto_descriptor = - getDescriptor().getMessageTypes().get(29); - internal_static_SetCapacityRequestProto_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_SetCapacityRequestProto_descriptor, - new java.lang.String[] { "ExecutorNum", "QueueSize", }); - internal_static_SetCapacityResponseProto_descriptor = - getDescriptor().getMessageTypes().get(30); - internal_static_SetCapacityResponseProto_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_SetCapacityResponseProto_descriptor, - new java.lang.String[] { }); - internal_static_EvictEntityRequestProto_descriptor = - getDescriptor().getMessageTypes().get(31); - internal_static_EvictEntityRequestProto_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_EvictEntityRequestProto_descriptor, - new java.lang.String[] { "DbName", "Table", }); - internal_static_TableProto_descriptor = - getDescriptor().getMessageTypes().get(32); - internal_static_TableProto_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_TableProto_descriptor, - new java.lang.String[] { "TableName", "PartKey", "PartVal", }); - internal_static_EvictEntityResponseProto_descriptor = - getDescriptor().getMessageTypes().get(33); - internal_static_EvictEntityResponseProto_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_EvictEntityResponseProto_descriptor, - new java.lang.String[] { "EvictedBytes", }); - internal_static_GetCacheContentRequestProto_descriptor = - getDescriptor().getMessageTypes().get(34); - internal_static_GetCacheContentRequestProto_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_GetCacheContentRequestProto_descriptor, - new java.lang.String[] { }); - internal_static_GetCacheContentResponseProto_descriptor = - getDescriptor().getMessageTypes().get(35); - internal_static_GetCacheContentResponseProto_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_GetCacheContentResponseProto_descriptor, - new java.lang.String[] { "Result", }); - internal_static_CacheEntryList_descriptor = - getDescriptor().getMessageTypes().get(36); - internal_static_CacheEntryList_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_CacheEntryList_descriptor, - new java.lang.String[] { "Entries", }); - internal_static_CacheEntry_descriptor = - getDescriptor().getMessageTypes().get(37); - internal_static_CacheEntry_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_CacheEntry_descriptor, - new java.lang.String[] { "FileKey", "FilePath", "CacheTag", "Ranges", }); - internal_static_CacheTag_descriptor = - getDescriptor().getMessageTypes().get(38); - internal_static_CacheTag_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_CacheTag_descriptor, - new java.lang.String[] { "TableName", "PartitionDesc", }); - internal_static_CacheEntryRange_descriptor = - getDescriptor().getMessageTypes().get(39); - internal_static_CacheEntryRange_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_CacheEntryRange_descriptor, - new java.lang.String[] { "Start", "End", }); - return null; - } - }; - com.google.protobuf.Descriptors.FileDescriptor + descriptor = com.google.protobuf.Descriptors.FileDescriptor .internalBuildGeneratedFileFrom(descriptorData, new com.google.protobuf.Descriptors.FileDescriptor[] { - }, assigner); + }); + internal_static_UserPayloadProto_descriptor = + getDescriptor().getMessageTypes().get(0); + internal_static_UserPayloadProto_fieldAccessorTable = new + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_UserPayloadProto_descriptor, + new java.lang.String[] { "UserPayload", "Version", }); + internal_static_EntityDescriptorProto_descriptor = + getDescriptor().getMessageTypes().get(1); + internal_static_EntityDescriptorProto_fieldAccessorTable = new + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_EntityDescriptorProto_descriptor, + new java.lang.String[] { "ClassName", "UserPayload", "HistoryText", }); + internal_static_IOSpecProto_descriptor = + getDescriptor().getMessageTypes().get(2); + internal_static_IOSpecProto_fieldAccessorTable = new + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_IOSpecProto_descriptor, + new java.lang.String[] { "ConnectedVertexName", "IoDescriptor", "PhysicalEdgeCount", }); + internal_static_GroupInputSpecProto_descriptor = + getDescriptor().getMessageTypes().get(3); + internal_static_GroupInputSpecProto_fieldAccessorTable = new + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_GroupInputSpecProto_descriptor, + new java.lang.String[] { "GroupName", "GroupVertices", "MergedInputDescriptor", }); + internal_static_SignableVertexSpec_descriptor = + getDescriptor().getMessageTypes().get(4); + internal_static_SignableVertexSpec_fieldAccessorTable = new + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_SignableVertexSpec_descriptor, + new java.lang.String[] { "User", "SignatureKeyId", "QueryIdentifier", "HiveQueryId", "DagName", "VertexName", "VertexIndex", "TokenIdentifier", "ProcessorDescriptor", "InputSpecs", "OutputSpecs", "GroupedInputSpecs", "VertexParallelism", "IsExternalSubmission", }); + internal_static_VertexOrBinary_descriptor = + getDescriptor().getMessageTypes().get(5); + internal_static_VertexOrBinary_fieldAccessorTable = new + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_VertexOrBinary_descriptor, + new java.lang.String[] { "Vertex", "VertexBinary", }); + internal_static_FragmentRuntimeInfo_descriptor = + getDescriptor().getMessageTypes().get(6); + internal_static_FragmentRuntimeInfo_fieldAccessorTable = new + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_FragmentRuntimeInfo_descriptor, + new java.lang.String[] { "NumSelfAndUpstreamTasks", "NumSelfAndUpstreamCompletedTasks", "WithinDagPriority", "DagStartTime", "FirstAttemptStartTime", "CurrentAttemptStartTime", }); + internal_static_QueryIdentifierProto_descriptor = + getDescriptor().getMessageTypes().get(7); + internal_static_QueryIdentifierProto_fieldAccessorTable = new + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_QueryIdentifierProto_descriptor, + new java.lang.String[] { "ApplicationIdString", "DagIndex", "AppAttemptNumber", }); + internal_static_NotTezEvent_descriptor = + getDescriptor().getMessageTypes().get(8); + internal_static_NotTezEvent_fieldAccessorTable = new + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_NotTezEvent_descriptor, + new java.lang.String[] { "InputEventProtoBytes", "VertexName", "DestInputName", "KeyId", }); + internal_static_SubmitWorkRequestProto_descriptor = + getDescriptor().getMessageTypes().get(9); + internal_static_SubmitWorkRequestProto_fieldAccessorTable = new + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_SubmitWorkRequestProto_descriptor, + new java.lang.String[] { "WorkSpec", "WorkSpecSignature", "FragmentNumber", "AttemptNumber", "ContainerIdString", "AmHost", "AmPort", "CredentialsBinary", "FragmentRuntimeInfo", "InitialEventBytes", "InitialEventSignature", "IsGuaranteed", "Jwt", "IsExternalClientRequest", }); + internal_static_RegisterDagRequestProto_descriptor = + getDescriptor().getMessageTypes().get(10); + internal_static_RegisterDagRequestProto_fieldAccessorTable = new + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_RegisterDagRequestProto_descriptor, + new java.lang.String[] { "User", "QueryIdentifier", "CredentialsBinary", }); + internal_static_RegisterDagResponseProto_descriptor = + getDescriptor().getMessageTypes().get(11); + internal_static_RegisterDagResponseProto_fieldAccessorTable = new + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_RegisterDagResponseProto_descriptor, + new java.lang.String[] { }); + internal_static_SubmitWorkResponseProto_descriptor = + getDescriptor().getMessageTypes().get(12); + internal_static_SubmitWorkResponseProto_fieldAccessorTable = new + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_SubmitWorkResponseProto_descriptor, + new java.lang.String[] { "SubmissionState", "UniqueNodeId", }); + internal_static_SourceStateUpdatedRequestProto_descriptor = + getDescriptor().getMessageTypes().get(13); + internal_static_SourceStateUpdatedRequestProto_fieldAccessorTable = new + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_SourceStateUpdatedRequestProto_descriptor, + new java.lang.String[] { "QueryIdentifier", "SrcName", "State", }); + internal_static_SourceStateUpdatedResponseProto_descriptor = + getDescriptor().getMessageTypes().get(14); + internal_static_SourceStateUpdatedResponseProto_fieldAccessorTable = new + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_SourceStateUpdatedResponseProto_descriptor, + new java.lang.String[] { }); + internal_static_QueryCompleteRequestProto_descriptor = + getDescriptor().getMessageTypes().get(15); + internal_static_QueryCompleteRequestProto_fieldAccessorTable = new + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_QueryCompleteRequestProto_descriptor, + new java.lang.String[] { "QueryIdentifier", "DeleteDelay", }); + internal_static_QueryCompleteResponseProto_descriptor = + getDescriptor().getMessageTypes().get(16); + internal_static_QueryCompleteResponseProto_fieldAccessorTable = new + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_QueryCompleteResponseProto_descriptor, + new java.lang.String[] { }); + internal_static_TerminateFragmentRequestProto_descriptor = + getDescriptor().getMessageTypes().get(17); + internal_static_TerminateFragmentRequestProto_fieldAccessorTable = new + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_TerminateFragmentRequestProto_descriptor, + new java.lang.String[] { "QueryIdentifier", "FragmentIdentifierString", }); + internal_static_TerminateFragmentResponseProto_descriptor = + getDescriptor().getMessageTypes().get(18); + internal_static_TerminateFragmentResponseProto_fieldAccessorTable = new + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_TerminateFragmentResponseProto_descriptor, + new java.lang.String[] { }); + internal_static_UpdateFragmentRequestProto_descriptor = + getDescriptor().getMessageTypes().get(19); + internal_static_UpdateFragmentRequestProto_fieldAccessorTable = new + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_UpdateFragmentRequestProto_descriptor, + new java.lang.String[] { "QueryIdentifier", "FragmentIdentifierString", "IsGuaranteed", }); + internal_static_UpdateFragmentResponseProto_descriptor = + getDescriptor().getMessageTypes().get(20); + internal_static_UpdateFragmentResponseProto_fieldAccessorTable = new + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_UpdateFragmentResponseProto_descriptor, + new java.lang.String[] { "Result", "IsGuaranteed", }); + internal_static_GetTokenRequestProto_descriptor = + getDescriptor().getMessageTypes().get(21); + internal_static_GetTokenRequestProto_fieldAccessorTable = new + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_GetTokenRequestProto_descriptor, + new java.lang.String[] { "AppId", }); + internal_static_GetTokenResponseProto_descriptor = + getDescriptor().getMessageTypes().get(22); + internal_static_GetTokenResponseProto_fieldAccessorTable = new + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_GetTokenResponseProto_descriptor, + new java.lang.String[] { "Token", }); + internal_static_LlapOutputSocketInitMessage_descriptor = + getDescriptor().getMessageTypes().get(23); + internal_static_LlapOutputSocketInitMessage_fieldAccessorTable = new + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_LlapOutputSocketInitMessage_descriptor, + new java.lang.String[] { "FragmentId", "Token", }); + internal_static_PurgeCacheRequestProto_descriptor = + getDescriptor().getMessageTypes().get(24); + internal_static_PurgeCacheRequestProto_fieldAccessorTable = new + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_PurgeCacheRequestProto_descriptor, + new java.lang.String[] { }); + internal_static_PurgeCacheResponseProto_descriptor = + getDescriptor().getMessageTypes().get(25); + internal_static_PurgeCacheResponseProto_fieldAccessorTable = new + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_PurgeCacheResponseProto_descriptor, + new java.lang.String[] { "PurgedMemoryBytes", }); + internal_static_MapEntry_descriptor = + getDescriptor().getMessageTypes().get(26); + internal_static_MapEntry_fieldAccessorTable = new + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_MapEntry_descriptor, + new java.lang.String[] { "Key", "Value", }); + internal_static_GetDaemonMetricsRequestProto_descriptor = + getDescriptor().getMessageTypes().get(27); + internal_static_GetDaemonMetricsRequestProto_fieldAccessorTable = new + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_GetDaemonMetricsRequestProto_descriptor, + new java.lang.String[] { }); + internal_static_GetDaemonMetricsResponseProto_descriptor = + getDescriptor().getMessageTypes().get(28); + internal_static_GetDaemonMetricsResponseProto_fieldAccessorTable = new + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_GetDaemonMetricsResponseProto_descriptor, + new java.lang.String[] { "Metrics", }); + internal_static_SetCapacityRequestProto_descriptor = + getDescriptor().getMessageTypes().get(29); + internal_static_SetCapacityRequestProto_fieldAccessorTable = new + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_SetCapacityRequestProto_descriptor, + new java.lang.String[] { "ExecutorNum", "QueueSize", }); + internal_static_SetCapacityResponseProto_descriptor = + getDescriptor().getMessageTypes().get(30); + internal_static_SetCapacityResponseProto_fieldAccessorTable = new + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_SetCapacityResponseProto_descriptor, + new java.lang.String[] { }); + internal_static_EvictEntityRequestProto_descriptor = + getDescriptor().getMessageTypes().get(31); + internal_static_EvictEntityRequestProto_fieldAccessorTable = new + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_EvictEntityRequestProto_descriptor, + new java.lang.String[] { "DbName", "Table", }); + internal_static_TableProto_descriptor = + getDescriptor().getMessageTypes().get(32); + internal_static_TableProto_fieldAccessorTable = new + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_TableProto_descriptor, + new java.lang.String[] { "TableName", "PartKey", "PartVal", }); + internal_static_EvictEntityResponseProto_descriptor = + getDescriptor().getMessageTypes().get(33); + internal_static_EvictEntityResponseProto_fieldAccessorTable = new + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_EvictEntityResponseProto_descriptor, + new java.lang.String[] { "EvictedBytes", }); + internal_static_GetCacheContentRequestProto_descriptor = + getDescriptor().getMessageTypes().get(34); + internal_static_GetCacheContentRequestProto_fieldAccessorTable = new + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_GetCacheContentRequestProto_descriptor, + new java.lang.String[] { }); + internal_static_GetCacheContentResponseProto_descriptor = + getDescriptor().getMessageTypes().get(35); + internal_static_GetCacheContentResponseProto_fieldAccessorTable = new + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_GetCacheContentResponseProto_descriptor, + new java.lang.String[] { "Result", }); + internal_static_CacheEntryList_descriptor = + getDescriptor().getMessageTypes().get(36); + internal_static_CacheEntryList_fieldAccessorTable = new + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_CacheEntryList_descriptor, + new java.lang.String[] { "Entries", }); + internal_static_CacheEntry_descriptor = + getDescriptor().getMessageTypes().get(37); + internal_static_CacheEntry_fieldAccessorTable = new + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_CacheEntry_descriptor, + new java.lang.String[] { "FileKey", "FilePath", "CacheTag", "Ranges", }); + internal_static_CacheTag_descriptor = + getDescriptor().getMessageTypes().get(38); + internal_static_CacheTag_fieldAccessorTable = new + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_CacheTag_descriptor, + new java.lang.String[] { "TableName", "PartitionDesc", }); + internal_static_CacheEntryRange_descriptor = + getDescriptor().getMessageTypes().get(39); + internal_static_CacheEntryRange_fieldAccessorTable = new + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_CacheEntryRange_descriptor, + new java.lang.String[] { "Start", "End", }); } // @@protoc_insertion_point(outer_class_scope) diff --git a/llap-common/src/gen/protobuf/gen-java/org/apache/hadoop/hive/llap/plugin/rpc/LlapPluginProtocolProtos.java b/llap-common/src/gen/protobuf/gen-java/org/apache/hadoop/hive/llap/plugin/rpc/LlapPluginProtocolProtos.java index dbcd895f77a0..97430d946204 100644 --- a/llap-common/src/gen/protobuf/gen-java/org/apache/hadoop/hive/llap/plugin/rpc/LlapPluginProtocolProtos.java +++ b/llap-common/src/gen/protobuf/gen-java/org/apache/hadoop/hive/llap/plugin/rpc/LlapPluginProtocolProtos.java @@ -5,19 +5,27 @@ public final class LlapPluginProtocolProtos { private LlapPluginProtocolProtos() {} + public static void registerAllExtensions( + com.google.protobuf.ExtensionRegistryLite registry) { + } + public static void registerAllExtensions( com.google.protobuf.ExtensionRegistry registry) { + registerAllExtensions( + (com.google.protobuf.ExtensionRegistryLite) registry); } - public interface UpdateQueryRequestProtoOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface UpdateQueryRequestProtoOrBuilder extends + // @@protoc_insertion_point(interface_extends:UpdateQueryRequestProto) + com.google.protobuf.MessageOrBuilder { - // optional int32 guaranteed_task_count = 1; /** * optional int32 guaranteed_task_count = 1; + * @return Whether the guaranteedTaskCount field is set. */ boolean hasGuaranteedTaskCount(); /** * optional int32 guaranteed_task_count = 1; + * @return The guaranteedTaskCount. */ int getGuaranteedTaskCount(); } @@ -25,35 +33,37 @@ public interface UpdateQueryRequestProtoOrBuilder * Protobuf type {@code UpdateQueryRequestProto} */ public static final class UpdateQueryRequestProto extends - com.google.protobuf.GeneratedMessage - implements UpdateQueryRequestProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:UpdateQueryRequestProto) + UpdateQueryRequestProtoOrBuilder { + private static final long serialVersionUID = 0L; // Use UpdateQueryRequestProto.newBuilder() to construct. - private UpdateQueryRequestProto(com.google.protobuf.GeneratedMessage.Builder builder) { + private UpdateQueryRequestProto(com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private UpdateQueryRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final UpdateQueryRequestProto defaultInstance; - public static UpdateQueryRequestProto getDefaultInstance() { - return defaultInstance; + private UpdateQueryRequestProto() { } - public UpdateQueryRequestProto getDefaultInstanceForType() { - return defaultInstance; + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance( + UnusedPrivateParameter unused) { + return new UpdateQueryRequestProto(); } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private UpdateQueryRequestProto( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); @@ -65,25 +75,25 @@ private UpdateQueryRequestProto( case 0: done = true; break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } case 8: { bitField0_ |= 0x00000001; guaranteedTaskCount_ = input.readInt32(); break; } + default: { + if (!parseUnknownField( + input, unknownFields, extensionRegistry, tag)) { + done = true; + } + break; + } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); @@ -94,88 +104,69 @@ private UpdateQueryRequestProto( return org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.internal_static_UpdateQueryRequestProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.internal_static_UpdateQueryRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryRequestProto.class, org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryRequestProto.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public UpdateQueryRequestProto parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new UpdateQueryRequestProto(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - private int bitField0_; - // optional int32 guaranteed_task_count = 1; public static final int GUARANTEED_TASK_COUNT_FIELD_NUMBER = 1; private int guaranteedTaskCount_; /** * optional int32 guaranteed_task_count = 1; + * @return Whether the guaranteedTaskCount field is set. */ + @java.lang.Override public boolean hasGuaranteedTaskCount() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional int32 guaranteed_task_count = 1; + * @return The guaranteedTaskCount. */ + @java.lang.Override public int getGuaranteedTaskCount() { return guaranteedTaskCount_; } - private void initFields() { - guaranteedTaskCount_ = 0; - } private byte memoizedIsInitialized = -1; + @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } + @java.lang.Override public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); - if (((bitField0_ & 0x00000001) == 0x00000001)) { + if (((bitField0_ & 0x00000001) != 0)) { output.writeInt32(1, guaranteedTaskCount_); } - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; + @java.lang.Override public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { + if (((bitField0_ & 0x00000001) != 0)) { size += com.google.protobuf.CodedOutputStream .computeInt32Size(1, guaranteedTaskCount_); } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { @@ -186,34 +177,42 @@ public boolean equals(final java.lang.Object obj) { } org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryRequestProto other = (org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryRequestProto) obj; - boolean result = true; - result = result && (hasGuaranteedTaskCount() == other.hasGuaranteedTaskCount()); + if (hasGuaranteedTaskCount() != other.hasGuaranteedTaskCount()) return false; if (hasGuaranteedTaskCount()) { - result = result && (getGuaranteedTaskCount() - == other.getGuaranteedTaskCount()); + if (getGuaranteedTaskCount() + != other.getGuaranteedTaskCount()) return false; } - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; + if (!unknownFields.equals(other.unknownFields)) return false; + return true; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); + hash = (19 * hash) + getDescriptor().hashCode(); if (hasGuaranteedTaskCount()) { hash = (37 * hash) + GUARANTEED_TASK_COUNT_FIELD_NUMBER; hash = (53 * hash) + getGuaranteedTaskCount(); } - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } + public static org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryRequestProto parseFrom( + java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryRequestProto parseFrom( + java.nio.ByteBuffer data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } public static org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryRequestProto parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { @@ -237,46 +236,59 @@ public static org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.Up } public static org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryRequestProto parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryRequestProto parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryRequestProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryRequestProto parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryRequestProto parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryRequestProto parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } + @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryRequestProto prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -284,14 +296,16 @@ protected Builder newBuilderForType( * Protobuf type {@code UpdateQueryRequestProto} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryRequestProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder implements + // @@protoc_insertion_point(builder_implements:UpdateQueryRequestProto) + org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryRequestProtoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.internal_static_UpdateQueryRequestProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.internal_static_UpdateQueryRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -304,18 +318,16 @@ private Builder() { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { } } - private static Builder create() { - return new Builder(); - } - + @java.lang.Override public Builder clear() { super.clear(); guaranteedTaskCount_ = 0; @@ -323,19 +335,18 @@ public Builder clear() { return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - + @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.internal_static_UpdateQueryRequestProto_descriptor; } + @java.lang.Override public org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryRequestProto getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryRequestProto.getDefaultInstance(); } + @java.lang.Override public org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryRequestProto build() { org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryRequestProto result = buildPartial(); if (!result.isInitialized()) { @@ -344,19 +355,53 @@ public org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQue return result; } + @java.lang.Override public org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryRequestProto buildPartial() { org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryRequestProto result = new org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryRequestProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { + if (((from_bitField0_ & 0x00000001) != 0)) { + result.guaranteedTaskCount_ = guaranteedTaskCount_; to_bitField0_ |= 0x00000001; } - result.guaranteedTaskCount_ = guaranteedTaskCount_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } + @java.lang.Override + public Builder clone() { + return super.clone(); + } + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.setField(field, value); + } + @java.lang.Override + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + @java.lang.Override + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.addRepeatedField(field, value); + } + @java.lang.Override public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryRequestProto) { return mergeFrom((org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryRequestProto)other); @@ -371,14 +416,17 @@ public Builder mergeFrom(org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtoc if (other.hasGuaranteedTaskCount()) { setGuaranteedTaskCount(other.getGuaranteedTaskCount()); } - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } + @java.lang.Override public final boolean isInitialized() { return true; } + @java.lang.Override public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) @@ -388,7 +436,7 @@ public Builder mergeFrom( parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryRequestProto) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -398,22 +446,27 @@ public Builder mergeFrom( } private int bitField0_; - // optional int32 guaranteed_task_count = 1; private int guaranteedTaskCount_ ; /** * optional int32 guaranteed_task_count = 1; + * @return Whether the guaranteedTaskCount field is set. */ + @java.lang.Override public boolean hasGuaranteedTaskCount() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional int32 guaranteed_task_count = 1; + * @return The guaranteedTaskCount. */ + @java.lang.Override public int getGuaranteedTaskCount() { return guaranteedTaskCount_; } /** * optional int32 guaranteed_task_count = 1; + * @param value The guaranteedTaskCount to set. + * @return This builder for chaining. */ public Builder setGuaranteedTaskCount(int value) { bitField0_ |= 0x00000001; @@ -423,6 +476,7 @@ public Builder setGuaranteedTaskCount(int value) { } /** * optional int32 guaranteed_task_count = 1; + * @return This builder for chaining. */ public Builder clearGuaranteedTaskCount() { bitField0_ = (bitField0_ & ~0x00000001); @@ -430,54 +484,98 @@ public Builder clearGuaranteedTaskCount() { onChanged(); return this; } + @java.lang.Override + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:UpdateQueryRequestProto) } + // @@protoc_insertion_point(class_scope:UpdateQueryRequestProto) + private static final org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryRequestProto DEFAULT_INSTANCE; static { - defaultInstance = new UpdateQueryRequestProto(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryRequestProto(); + } + + public static org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryRequestProto getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser + PARSER = new com.google.protobuf.AbstractParser() { + @java.lang.Override + public UpdateQueryRequestProto parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new UpdateQueryRequestProto(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryRequestProto getDefaultInstanceForType() { + return DEFAULT_INSTANCE; } - // @@protoc_insertion_point(class_scope:UpdateQueryRequestProto) } - public interface UpdateQueryResponseProtoOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface UpdateQueryResponseProtoOrBuilder extends + // @@protoc_insertion_point(interface_extends:UpdateQueryResponseProto) + com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code UpdateQueryResponseProto} */ public static final class UpdateQueryResponseProto extends - com.google.protobuf.GeneratedMessage - implements UpdateQueryResponseProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:UpdateQueryResponseProto) + UpdateQueryResponseProtoOrBuilder { + private static final long serialVersionUID = 0L; // Use UpdateQueryResponseProto.newBuilder() to construct. - private UpdateQueryResponseProto(com.google.protobuf.GeneratedMessage.Builder builder) { + private UpdateQueryResponseProto(com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private UpdateQueryResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final UpdateQueryResponseProto defaultInstance; - public static UpdateQueryResponseProto getDefaultInstance() { - return defaultInstance; + private UpdateQueryResponseProto() { } - public UpdateQueryResponseProto getDefaultInstanceForType() { - return defaultInstance; + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance( + UnusedPrivateParameter unused) { + return new UpdateQueryResponseProto(); } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private UpdateQueryResponseProto( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); try { @@ -489,8 +587,8 @@ private UpdateQueryResponseProto( done = true; break; default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { + if (!parseUnknownField( + input, unknownFields, extensionRegistry, tag)) { done = true; } break; @@ -501,7 +599,7 @@ private UpdateQueryResponseProto( throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); @@ -512,63 +610,42 @@ private UpdateQueryResponseProto( return org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.internal_static_UpdateQueryResponseProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.internal_static_UpdateQueryResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryResponseProto.class, org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryResponseProto.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public UpdateQueryResponseProto parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new UpdateQueryResponseProto(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - - private void initFields() { - } private byte memoizedIsInitialized = -1; + @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } + @java.lang.Override public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; + @java.lang.Override public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } - private static final long serialVersionUID = 0L; - @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); - } - @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { @@ -579,25 +656,33 @@ public boolean equals(final java.lang.Object obj) { } org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryResponseProto other = (org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryResponseProto) obj; - boolean result = true; - result = result && - getUnknownFields().equals(other.getUnknownFields()); - return result; + if (!unknownFields.equals(other.unknownFields)) return false; + return true; } - private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; - hash = (19 * hash) + getDescriptorForType().hashCode(); - hash = (29 * hash) + getUnknownFields().hashCode(); + hash = (19 * hash) + getDescriptor().hashCode(); + hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } + public static org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryResponseProto parseFrom( + java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryResponseProto parseFrom( + java.nio.ByteBuffer data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } public static org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryResponseProto parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { @@ -621,46 +706,59 @@ public static org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.Up } public static org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryResponseProto parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryResponseProto parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryResponseProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryResponseProto parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryResponseProto parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryResponseProto parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } + @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryResponseProto prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -668,14 +766,16 @@ protected Builder newBuilderForType( * Protobuf type {@code UpdateQueryResponseProto} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryResponseProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder implements + // @@protoc_insertion_point(builder_implements:UpdateQueryResponseProto) + org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryResponseProtoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.internal_static_UpdateQueryResponseProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.internal_static_UpdateQueryResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -688,36 +788,33 @@ private Builder() { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { } } - private static Builder create() { - return new Builder(); - } - + @java.lang.Override public Builder clear() { super.clear(); return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - + @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.internal_static_UpdateQueryResponseProto_descriptor; } + @java.lang.Override public org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryResponseProto getDefaultInstanceForType() { return org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryResponseProto.getDefaultInstance(); } + @java.lang.Override public org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryResponseProto build() { org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryResponseProto result = buildPartial(); if (!result.isInitialized()) { @@ -726,12 +823,46 @@ public org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQue return result; } + @java.lang.Override public org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryResponseProto buildPartial() { org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryResponseProto result = new org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryResponseProto(this); onBuilt(); return result; } + @java.lang.Override + public Builder clone() { + return super.clone(); + } + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.setField(field, value); + } + @java.lang.Override + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + @java.lang.Override + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.addRepeatedField(field, value); + } + @java.lang.Override public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryResponseProto) { return mergeFrom((org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryResponseProto)other); @@ -743,14 +874,17 @@ public Builder mergeFrom(com.google.protobuf.Message other) { public Builder mergeFrom(org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryResponseProto other) { if (other == org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryResponseProto.getDefaultInstance()) return this; - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } + @java.lang.Override public final boolean isInitialized() { return true; } + @java.lang.Override public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) @@ -760,7 +894,7 @@ public Builder mergeFrom( parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryResponseProto) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -768,16 +902,57 @@ public Builder mergeFrom( } return this; } + @java.lang.Override + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:UpdateQueryResponseProto) } + // @@protoc_insertion_point(class_scope:UpdateQueryResponseProto) + private static final org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryResponseProto DEFAULT_INSTANCE; static { - defaultInstance = new UpdateQueryResponseProto(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryResponseProto(); + } + + public static org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryResponseProto getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser + PARSER = new com.google.protobuf.AbstractParser() { + @java.lang.Override + public UpdateQueryResponseProto parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new UpdateQueryResponseProto(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQueryResponseProto getDefaultInstanceForType() { + return DEFAULT_INSTANCE; } - // @@protoc_insertion_point(class_scope:UpdateQueryResponseProto) } /** @@ -1012,22 +1187,22 @@ public org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQue // @@protoc_insertion_point(class_scope:LlapPluginProtocol) } - private static com.google.protobuf.Descriptors.Descriptor + private static final com.google.protobuf.Descriptors.Descriptor internal_static_UpdateQueryRequestProto_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + private static final + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_UpdateQueryRequestProto_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static final com.google.protobuf.Descriptors.Descriptor internal_static_UpdateQueryResponseProto_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + private static final + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_UpdateQueryResponseProto_fieldAccessorTable; public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { return descriptor; } - private static com.google.protobuf.Descriptors.FileDescriptor + private static com.google.protobuf.Descriptors.FileDescriptor descriptor; static { java.lang.String[] descriptorData = { @@ -1039,30 +1214,22 @@ public org.apache.hadoop.hive.llap.plugin.rpc.LlapPluginProtocolProtos.UpdateQue "toBH\n&org.apache.hadoop.hive.llap.plugin" + ".rpcB\030LlapPluginProtocolProtos\210\001\001\240\001\001" }; - com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = - new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { - public com.google.protobuf.ExtensionRegistry assignDescriptors( - com.google.protobuf.Descriptors.FileDescriptor root) { - descriptor = root; - internal_static_UpdateQueryRequestProto_descriptor = - getDescriptor().getMessageTypes().get(0); - internal_static_UpdateQueryRequestProto_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_UpdateQueryRequestProto_descriptor, - new java.lang.String[] { "GuaranteedTaskCount", }); - internal_static_UpdateQueryResponseProto_descriptor = - getDescriptor().getMessageTypes().get(1); - internal_static_UpdateQueryResponseProto_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_UpdateQueryResponseProto_descriptor, - new java.lang.String[] { }); - return null; - } - }; - com.google.protobuf.Descriptors.FileDescriptor + descriptor = com.google.protobuf.Descriptors.FileDescriptor .internalBuildGeneratedFileFrom(descriptorData, new com.google.protobuf.Descriptors.FileDescriptor[] { - }, assigner); + }); + internal_static_UpdateQueryRequestProto_descriptor = + getDescriptor().getMessageTypes().get(0); + internal_static_UpdateQueryRequestProto_fieldAccessorTable = new + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_UpdateQueryRequestProto_descriptor, + new java.lang.String[] { "GuaranteedTaskCount", }); + internal_static_UpdateQueryResponseProto_descriptor = + getDescriptor().getMessageTypes().get(1); + internal_static_UpdateQueryResponseProto_fieldAccessorTable = new + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_UpdateQueryResponseProto_descriptor, + new java.lang.String[] { }); } // @@protoc_insertion_point(outer_class_scope) diff --git a/pom.xml b/pom.xml index bc4e5fa1d83c..403f5c5c43fd 100644 --- a/pom.xml +++ b/pom.xml @@ -132,7 +132,7 @@ 3.1.0 ${basedir}/${hive.path.to.root}/testutils/hadoop 1.3 - 2.0.0-alpha4 + 2.2.7 0.7.2 3.3.7 @@ -185,7 +185,7 @@ 1.11.1 0.16.0 1.5.6 - 2.5.0 + 3.21.1 1.0.1 1.7.30 4.0.4 @@ -193,7 +193,7 @@ 0.10.1 2.2.0 1.1 - 1.1.4 + 1.1.8.4 1.4 2.3 2.12.2 @@ -1135,6 +1135,11 @@
+ + org.apache.hbase + hbase-shaded-client-byo-hadoop + ${hbase.version} + org.apache.hbase hbase-server @@ -1517,7 +1522,6 @@ Do not use shaded imports - **.shaded.** jersey.repackaged.com.google.** org.codehaus.jackson.** diff --git a/ql/src/gen/protobuf/gen-java/org/apache/hadoop/hive/ql/hooks/proto/HiveHookEvents.java b/ql/src/gen/protobuf/gen-java/org/apache/hadoop/hive/ql/hooks/proto/HiveHookEvents.java index 6adda7283ade..d3770f65aa07 100644 --- a/ql/src/gen/protobuf/gen-java/org/apache/hadoop/hive/ql/hooks/proto/HiveHookEvents.java +++ b/ql/src/gen/protobuf/gen-java/org/apache/hadoop/hive/ql/hooks/proto/HiveHookEvents.java @@ -5,75 +5,90 @@ public final class HiveHookEvents { private HiveHookEvents() {} + public static void registerAllExtensions( + com.google.protobuf.ExtensionRegistryLite registry) { + } + public static void registerAllExtensions( com.google.protobuf.ExtensionRegistry registry) { + registerAllExtensions( + (com.google.protobuf.ExtensionRegistryLite) registry); } - public interface MapFieldEntryOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface MapFieldEntryOrBuilder extends + // @@protoc_insertion_point(interface_extends:MapFieldEntry) + com.google.protobuf.MessageOrBuilder { - // optional string key = 1; /** * optional string key = 1; + * @return Whether the key field is set. */ boolean hasKey(); /** * optional string key = 1; + * @return The key. */ java.lang.String getKey(); /** * optional string key = 1; + * @return The bytes for key. */ com.google.protobuf.ByteString getKeyBytes(); - // optional string value = 2; /** * optional string value = 2; + * @return Whether the value field is set. */ boolean hasValue(); /** * optional string value = 2; + * @return The value. */ java.lang.String getValue(); /** * optional string value = 2; + * @return The bytes for value. */ com.google.protobuf.ByteString getValueBytes(); } /** - * Protobuf type {@code org.apache.hadoop.hive.ql.hooks.proto.MapFieldEntry} + * Protobuf type {@code MapFieldEntry} */ public static final class MapFieldEntry extends - com.google.protobuf.GeneratedMessage - implements MapFieldEntryOrBuilder { + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:MapFieldEntry) + MapFieldEntryOrBuilder { + private static final long serialVersionUID = 0L; // Use MapFieldEntry.newBuilder() to construct. - private MapFieldEntry(com.google.protobuf.GeneratedMessage.Builder builder) { + private MapFieldEntry(com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private MapFieldEntry(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final MapFieldEntry defaultInstance; - public static MapFieldEntry getDefaultInstance() { - return defaultInstance; + private MapFieldEntry() { + key_ = ""; + value_ = ""; } - public MapFieldEntry getDefaultInstanceForType() { - return defaultInstance; + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance( + UnusedPrivateParameter unused) { + return new MapFieldEntry(); } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private MapFieldEntry( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); @@ -85,21 +100,23 @@ private MapFieldEntry( case 0: done = true; break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } case 10: { + com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000001; - key_ = input.readBytes(); + key_ = bs; break; } case 18: { + com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000002; - value_ = input.readBytes(); + value_ = bs; + break; + } + default: { + if (!parseUnknownField( + input, unknownFields, extensionRegistry, tag)) { + done = true; + } break; } } @@ -108,7 +125,7 @@ private MapFieldEntry( throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); @@ -116,44 +133,33 @@ private MapFieldEntry( } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { - return org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.internal_static_org_apache_hadoop_hive_ql_hooks_proto_MapFieldEntry_descriptor; + return org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.internal_static_MapFieldEntry_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { - return org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.internal_static_org_apache_hadoop_hive_ql_hooks_proto_MapFieldEntry_fieldAccessorTable + return org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.internal_static_MapFieldEntry_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry.class, org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public MapFieldEntry parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new MapFieldEntry(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - private int bitField0_; - // optional string key = 1; public static final int KEY_FIELD_NUMBER = 1; - private java.lang.Object key_; + private volatile java.lang.Object key_; /** * optional string key = 1; + * @return Whether the key field is set. */ + @java.lang.Override public boolean hasKey() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional string key = 1; + * @return The key. */ + @java.lang.Override public java.lang.String getKey() { java.lang.Object ref = key_; if (ref instanceof java.lang.String) { @@ -170,7 +176,9 @@ public java.lang.String getKey() { } /** * optional string key = 1; + * @return The bytes for key. */ + @java.lang.Override public com.google.protobuf.ByteString getKeyBytes() { java.lang.Object ref = key_; @@ -185,18 +193,21 @@ public java.lang.String getKey() { } } - // optional string value = 2; public static final int VALUE_FIELD_NUMBER = 2; - private java.lang.Object value_; + private volatile java.lang.Object value_; /** * optional string value = 2; + * @return Whether the value field is set. */ + @java.lang.Override public boolean hasValue() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * optional string value = 2; + * @return The value. */ + @java.lang.Override public java.lang.String getValue() { java.lang.Object ref = value_; if (ref instanceof java.lang.String) { @@ -213,7 +224,9 @@ public java.lang.String getValue() { } /** * optional string value = 2; + * @return The bytes for value. */ + @java.lang.Override public com.google.protobuf.ByteString getValueBytes() { java.lang.Object ref = value_; @@ -228,57 +241,101 @@ public java.lang.String getValue() { } } - private void initFields() { - key_ = ""; - value_ = ""; - } private byte memoizedIsInitialized = -1; + @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } + @java.lang.Override public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); - if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeBytes(1, getKeyBytes()); + if (((bitField0_ & 0x00000001) != 0)) { + com.google.protobuf.GeneratedMessageV3.writeString(output, 1, key_); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - output.writeBytes(2, getValueBytes()); + if (((bitField0_ & 0x00000002) != 0)) { + com.google.protobuf.GeneratedMessageV3.writeString(output, 2, value_); } - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; + @java.lang.Override public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(1, getKeyBytes()); + if (((bitField0_ & 0x00000001) != 0)) { + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, key_); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(2, getValueBytes()); + if (((bitField0_ & 0x00000002) != 0)) { + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(2, value_); } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } - private static final long serialVersionUID = 0L; @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); + public boolean equals(final java.lang.Object obj) { + if (obj == this) { + return true; + } + if (!(obj instanceof org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry)) { + return super.equals(obj); + } + org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry other = (org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry) obj; + + if (hasKey() != other.hasKey()) return false; + if (hasKey()) { + if (!getKey() + .equals(other.getKey())) return false; + } + if (hasValue() != other.hasValue()) return false; + if (hasValue()) { + if (!getValue() + .equals(other.getValue())) return false; + } + if (!unknownFields.equals(other.unknownFields)) return false; + return true; + } + + @java.lang.Override + public int hashCode() { + if (memoizedHashCode != 0) { + return memoizedHashCode; + } + int hash = 41; + hash = (19 * hash) + getDescriptor().hashCode(); + if (hasKey()) { + hash = (37 * hash) + KEY_FIELD_NUMBER; + hash = (53 * hash) + getKey().hashCode(); + } + if (hasValue()) { + hash = (37 * hash) + VALUE_FIELD_NUMBER; + hash = (53 * hash) + getValue().hashCode(); + } + hash = (29 * hash) + unknownFields.hashCode(); + memoizedHashCode = hash; + return hash; } + public static org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry parseFrom( + java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry parseFrom( + java.nio.ByteBuffer data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } public static org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { @@ -302,63 +359,78 @@ public static org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry } public static org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } + @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** - * Protobuf type {@code org.apache.hadoop.hive.ql.hooks.proto.MapFieldEntry} + * Protobuf type {@code MapFieldEntry} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntryOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder implements + // @@protoc_insertion_point(builder_implements:MapFieldEntry) + org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntryOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { - return org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.internal_static_org_apache_hadoop_hive_ql_hooks_proto_MapFieldEntry_descriptor; + return org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.internal_static_MapFieldEntry_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { - return org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.internal_static_org_apache_hadoop_hive_ql_hooks_proto_MapFieldEntry_fieldAccessorTable + return org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.internal_static_MapFieldEntry_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry.class, org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry.Builder.class); } @@ -369,18 +441,16 @@ private Builder() { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { } } - private static Builder create() { - return new Builder(); - } - + @java.lang.Override public Builder clear() { super.clear(); key_ = ""; @@ -390,19 +460,18 @@ public Builder clear() { return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - + @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { - return org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.internal_static_org_apache_hadoop_hive_ql_hooks_proto_MapFieldEntry_descriptor; + return org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.internal_static_MapFieldEntry_descriptor; } + @java.lang.Override public org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry getDefaultInstanceForType() { return org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry.getDefaultInstance(); } + @java.lang.Override public org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry build() { org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry result = buildPartial(); if (!result.isInitialized()) { @@ -411,15 +480,16 @@ public org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry build( return result; } + @java.lang.Override public org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry buildPartial() { org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry result = new org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { + if (((from_bitField0_ & 0x00000001) != 0)) { to_bitField0_ |= 0x00000001; } result.key_ = key_; - if (((from_bitField0_ & 0x00000002) == 0x00000002)) { + if (((from_bitField0_ & 0x00000002) != 0)) { to_bitField0_ |= 0x00000002; } result.value_ = value_; @@ -428,6 +498,39 @@ public org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry buildP return result; } + @java.lang.Override + public Builder clone() { + return super.clone(); + } + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.setField(field, value); + } + @java.lang.Override + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + @java.lang.Override + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.addRepeatedField(field, value); + } + @java.lang.Override public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry) { return mergeFrom((org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry)other); @@ -449,14 +552,17 @@ public Builder mergeFrom(org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.Ma value_ = other.value_; onChanged(); } - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } + @java.lang.Override public final boolean isInitialized() { return true; } + @java.lang.Override public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) @@ -466,7 +572,7 @@ public Builder mergeFrom( parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -476,23 +582,27 @@ public Builder mergeFrom( } private int bitField0_; - // optional string key = 1; private java.lang.Object key_ = ""; /** * optional string key = 1; + * @return Whether the key field is set. */ public boolean hasKey() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional string key = 1; + * @return The key. */ public java.lang.String getKey() { java.lang.Object ref = key_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - key_ = s; + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + key_ = s; + } return s; } else { return (java.lang.String) ref; @@ -500,6 +610,7 @@ public java.lang.String getKey() { } /** * optional string key = 1; + * @return The bytes for key. */ public com.google.protobuf.ByteString getKeyBytes() { @@ -516,6 +627,8 @@ public java.lang.String getKey() { } /** * optional string key = 1; + * @param value The key to set. + * @return This builder for chaining. */ public Builder setKey( java.lang.String value) { @@ -529,6 +642,7 @@ public Builder setKey( } /** * optional string key = 1; + * @return This builder for chaining. */ public Builder clearKey() { bitField0_ = (bitField0_ & ~0x00000001); @@ -538,6 +652,8 @@ public Builder clearKey() { } /** * optional string key = 1; + * @param value The bytes for key to set. + * @return This builder for chaining. */ public Builder setKeyBytes( com.google.protobuf.ByteString value) { @@ -550,23 +666,27 @@ public Builder setKeyBytes( return this; } - // optional string value = 2; private java.lang.Object value_ = ""; /** * optional string value = 2; + * @return Whether the value field is set. */ public boolean hasValue() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * optional string value = 2; + * @return The value. */ public java.lang.String getValue() { java.lang.Object ref = value_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - value_ = s; + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + value_ = s; + } return s; } else { return (java.lang.String) ref; @@ -574,6 +694,7 @@ public java.lang.String getValue() { } /** * optional string value = 2; + * @return The bytes for value. */ public com.google.protobuf.ByteString getValueBytes() { @@ -590,6 +711,8 @@ public java.lang.String getValue() { } /** * optional string value = 2; + * @param value The value to set. + * @return This builder for chaining. */ public Builder setValue( java.lang.String value) { @@ -603,6 +726,7 @@ public Builder setValue( } /** * optional string value = 2; + * @return This builder for chaining. */ public Builder clearValue() { bitField0_ = (bitField0_ & ~0x00000002); @@ -612,6 +736,8 @@ public Builder clearValue() { } /** * optional string value = 2; + * @param value The bytes for value to set. + * @return This builder for chaining. */ public Builder setValueBytes( com.google.protobuf.ByteString value) { @@ -623,234 +749,312 @@ public Builder setValueBytes( onChanged(); return this; } + @java.lang.Override + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + - // @@protoc_insertion_point(builder_scope:org.apache.hadoop.hive.ql.hooks.proto.MapFieldEntry) + // @@protoc_insertion_point(builder_scope:MapFieldEntry) } + // @@protoc_insertion_point(class_scope:MapFieldEntry) + private static final org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry DEFAULT_INSTANCE; static { - defaultInstance = new MapFieldEntry(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry(); + } + + public static org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser + PARSER = new com.google.protobuf.AbstractParser() { + @java.lang.Override + public MapFieldEntry parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new MapFieldEntry(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry getDefaultInstanceForType() { + return DEFAULT_INSTANCE; } - // @@protoc_insertion_point(class_scope:org.apache.hadoop.hive.ql.hooks.proto.MapFieldEntry) } - public interface HiveHookEventProtoOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface HiveHookEventProtoOrBuilder extends + // @@protoc_insertion_point(interface_extends:HiveHookEventProto) + com.google.protobuf.MessageOrBuilder { - // optional string eventType = 1; /** * optional string eventType = 1; + * @return Whether the eventType field is set. */ boolean hasEventType(); /** * optional string eventType = 1; + * @return The eventType. */ java.lang.String getEventType(); /** * optional string eventType = 1; + * @return The bytes for eventType. */ com.google.protobuf.ByteString getEventTypeBytes(); - // optional string hiveQueryId = 2; /** * optional string hiveQueryId = 2; + * @return Whether the hiveQueryId field is set. */ boolean hasHiveQueryId(); /** * optional string hiveQueryId = 2; + * @return The hiveQueryId. */ java.lang.String getHiveQueryId(); /** * optional string hiveQueryId = 2; + * @return The bytes for hiveQueryId. */ com.google.protobuf.ByteString getHiveQueryIdBytes(); - // optional int64 timestamp = 3; /** * optional int64 timestamp = 3; + * @return Whether the timestamp field is set. */ boolean hasTimestamp(); /** * optional int64 timestamp = 3; + * @return The timestamp. */ long getTimestamp(); - // optional string executionMode = 4; /** * optional string executionMode = 4; + * @return Whether the executionMode field is set. */ boolean hasExecutionMode(); /** * optional string executionMode = 4; + * @return The executionMode. */ java.lang.String getExecutionMode(); /** * optional string executionMode = 4; + * @return The bytes for executionMode. */ com.google.protobuf.ByteString getExecutionModeBytes(); - // optional string requestUser = 5; /** * optional string requestUser = 5; + * @return Whether the requestUser field is set. */ boolean hasRequestUser(); /** * optional string requestUser = 5; + * @return The requestUser. */ java.lang.String getRequestUser(); /** * optional string requestUser = 5; + * @return The bytes for requestUser. */ com.google.protobuf.ByteString getRequestUserBytes(); - // optional string queue = 6; /** * optional string queue = 6; + * @return Whether the queue field is set. */ boolean hasQueue(); /** * optional string queue = 6; + * @return The queue. */ java.lang.String getQueue(); /** * optional string queue = 6; + * @return The bytes for queue. */ com.google.protobuf.ByteString getQueueBytes(); - // optional string user = 7; /** * optional string user = 7; + * @return Whether the user field is set. */ boolean hasUser(); /** * optional string user = 7; + * @return The user. */ java.lang.String getUser(); /** * optional string user = 7; + * @return The bytes for user. */ com.google.protobuf.ByteString getUserBytes(); - // optional string operationId = 8; /** * optional string operationId = 8; + * @return Whether the operationId field is set. */ boolean hasOperationId(); /** * optional string operationId = 8; + * @return The operationId. */ java.lang.String getOperationId(); /** * optional string operationId = 8; + * @return The bytes for operationId. */ com.google.protobuf.ByteString getOperationIdBytes(); - // repeated string tablesWritten = 9; /** * repeated string tablesWritten = 9; + * @return A list containing the tablesWritten. */ java.util.List - getTablesWrittenList(); + getTablesWrittenList(); /** * repeated string tablesWritten = 9; + * @return The count of tablesWritten. */ int getTablesWrittenCount(); /** * repeated string tablesWritten = 9; + * @param index The index of the element to return. + * @return The tablesWritten at the given index. */ java.lang.String getTablesWritten(int index); /** * repeated string tablesWritten = 9; + * @param index The index of the value to return. + * @return The bytes of the tablesWritten at the given index. */ com.google.protobuf.ByteString getTablesWrittenBytes(int index); - // repeated string tablesRead = 10; /** * repeated string tablesRead = 10; + * @return A list containing the tablesRead. */ java.util.List - getTablesReadList(); + getTablesReadList(); /** * repeated string tablesRead = 10; + * @return The count of tablesRead. */ int getTablesReadCount(); /** * repeated string tablesRead = 10; + * @param index The index of the element to return. + * @return The tablesRead at the given index. */ java.lang.String getTablesRead(int index); /** * repeated string tablesRead = 10; + * @param index The index of the value to return. + * @return The bytes of the tablesRead at the given index. */ com.google.protobuf.ByteString getTablesReadBytes(int index); - // repeated .org.apache.hadoop.hive.ql.hooks.proto.MapFieldEntry otherInfo = 50; /** - * repeated .org.apache.hadoop.hive.ql.hooks.proto.MapFieldEntry otherInfo = 50; + * repeated .MapFieldEntry otherInfo = 50; */ java.util.List getOtherInfoList(); /** - * repeated .org.apache.hadoop.hive.ql.hooks.proto.MapFieldEntry otherInfo = 50; + * repeated .MapFieldEntry otherInfo = 50; */ org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry getOtherInfo(int index); /** - * repeated .org.apache.hadoop.hive.ql.hooks.proto.MapFieldEntry otherInfo = 50; + * repeated .MapFieldEntry otherInfo = 50; */ int getOtherInfoCount(); /** - * repeated .org.apache.hadoop.hive.ql.hooks.proto.MapFieldEntry otherInfo = 50; + * repeated .MapFieldEntry otherInfo = 50; */ java.util.List getOtherInfoOrBuilderList(); /** - * repeated .org.apache.hadoop.hive.ql.hooks.proto.MapFieldEntry otherInfo = 50; + * repeated .MapFieldEntry otherInfo = 50; */ org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntryOrBuilder getOtherInfoOrBuilder( int index); } /** - * Protobuf type {@code org.apache.hadoop.hive.ql.hooks.proto.HiveHookEventProto} + * Protobuf type {@code HiveHookEventProto} */ public static final class HiveHookEventProto extends - com.google.protobuf.GeneratedMessage - implements HiveHookEventProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:HiveHookEventProto) + HiveHookEventProtoOrBuilder { + private static final long serialVersionUID = 0L; // Use HiveHookEventProto.newBuilder() to construct. - private HiveHookEventProto(com.google.protobuf.GeneratedMessage.Builder builder) { + private HiveHookEventProto(com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private HiveHookEventProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final HiveHookEventProto defaultInstance; - public static HiveHookEventProto getDefaultInstance() { - return defaultInstance; + private HiveHookEventProto() { + eventType_ = ""; + hiveQueryId_ = ""; + executionMode_ = ""; + requestUser_ = ""; + queue_ = ""; + user_ = ""; + operationId_ = ""; + tablesWritten_ = com.google.protobuf.LazyStringArrayList.EMPTY; + tablesRead_ = com.google.protobuf.LazyStringArrayList.EMPTY; + otherInfo_ = java.util.Collections.emptyList(); } - public HiveHookEventProto getDefaultInstanceForType() { - return defaultInstance; + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance( + UnusedPrivateParameter unused) { + return new HiveHookEventProto(); } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private HiveHookEventProto( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); @@ -862,21 +1066,16 @@ private HiveHookEventProto( case 0: done = true; break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } case 10: { + com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000001; - eventType_ = input.readBytes(); + eventType_ = bs; break; } case 18: { + com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000002; - hiveQueryId_ = input.readBytes(); + hiveQueryId_ = bs; break; } case 24: { @@ -885,52 +1084,67 @@ private HiveHookEventProto( break; } case 34: { + com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000008; - executionMode_ = input.readBytes(); + executionMode_ = bs; break; } case 42: { + com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000010; - requestUser_ = input.readBytes(); + requestUser_ = bs; break; } case 50: { + com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000020; - queue_ = input.readBytes(); + queue_ = bs; break; } case 58: { + com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000040; - user_ = input.readBytes(); + user_ = bs; break; } case 66: { + com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000080; - operationId_ = input.readBytes(); + operationId_ = bs; break; } case 74: { - if (!((mutable_bitField0_ & 0x00000100) == 0x00000100)) { + com.google.protobuf.ByteString bs = input.readBytes(); + if (!((mutable_bitField0_ & 0x00000100) != 0)) { tablesWritten_ = new com.google.protobuf.LazyStringArrayList(); mutable_bitField0_ |= 0x00000100; } - tablesWritten_.add(input.readBytes()); + tablesWritten_.add(bs); break; } case 82: { - if (!((mutable_bitField0_ & 0x00000200) == 0x00000200)) { + com.google.protobuf.ByteString bs = input.readBytes(); + if (!((mutable_bitField0_ & 0x00000200) != 0)) { tablesRead_ = new com.google.protobuf.LazyStringArrayList(); mutable_bitField0_ |= 0x00000200; } - tablesRead_.add(input.readBytes()); + tablesRead_.add(bs); break; } case 402: { - if (!((mutable_bitField0_ & 0x00000400) == 0x00000400)) { + if (!((mutable_bitField0_ & 0x00000400) != 0)) { otherInfo_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000400; } - otherInfo_.add(input.readMessage(org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry.PARSER, extensionRegistry)); + otherInfo_.add( + input.readMessage(org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry.PARSER, extensionRegistry)); + break; + } + default: { + if (!parseUnknownField( + input, unknownFields, extensionRegistry, tag)) { + done = true; + } break; } } @@ -939,15 +1153,15 @@ private HiveHookEventProto( throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { - if (((mutable_bitField0_ & 0x00000100) == 0x00000100)) { - tablesWritten_ = new com.google.protobuf.UnmodifiableLazyStringList(tablesWritten_); + if (((mutable_bitField0_ & 0x00000100) != 0)) { + tablesWritten_ = tablesWritten_.getUnmodifiableView(); } - if (((mutable_bitField0_ & 0x00000200) == 0x00000200)) { - tablesRead_ = new com.google.protobuf.UnmodifiableLazyStringList(tablesRead_); + if (((mutable_bitField0_ & 0x00000200) != 0)) { + tablesRead_ = tablesRead_.getUnmodifiableView(); } - if (((mutable_bitField0_ & 0x00000400) == 0x00000400)) { + if (((mutable_bitField0_ & 0x00000400) != 0)) { otherInfo_ = java.util.Collections.unmodifiableList(otherInfo_); } this.unknownFields = unknownFields.build(); @@ -956,44 +1170,33 @@ private HiveHookEventProto( } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { - return org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.internal_static_org_apache_hadoop_hive_ql_hooks_proto_HiveHookEventProto_descriptor; + return org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.internal_static_HiveHookEventProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { - return org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.internal_static_org_apache_hadoop_hive_ql_hooks_proto_HiveHookEventProto_fieldAccessorTable + return org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.internal_static_HiveHookEventProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.HiveHookEventProto.class, org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.HiveHookEventProto.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public HiveHookEventProto parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new HiveHookEventProto(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - private int bitField0_; - // optional string eventType = 1; public static final int EVENTTYPE_FIELD_NUMBER = 1; - private java.lang.Object eventType_; + private volatile java.lang.Object eventType_; /** * optional string eventType = 1; + * @return Whether the eventType field is set. */ + @java.lang.Override public boolean hasEventType() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional string eventType = 1; + * @return The eventType. */ + @java.lang.Override public java.lang.String getEventType() { java.lang.Object ref = eventType_; if (ref instanceof java.lang.String) { @@ -1010,7 +1213,9 @@ public java.lang.String getEventType() { } /** * optional string eventType = 1; + * @return The bytes for eventType. */ + @java.lang.Override public com.google.protobuf.ByteString getEventTypeBytes() { java.lang.Object ref = eventType_; @@ -1025,18 +1230,21 @@ public java.lang.String getEventType() { } } - // optional string hiveQueryId = 2; public static final int HIVEQUERYID_FIELD_NUMBER = 2; - private java.lang.Object hiveQueryId_; + private volatile java.lang.Object hiveQueryId_; /** * optional string hiveQueryId = 2; + * @return Whether the hiveQueryId field is set. */ + @java.lang.Override public boolean hasHiveQueryId() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * optional string hiveQueryId = 2; + * @return The hiveQueryId. */ + @java.lang.Override public java.lang.String getHiveQueryId() { java.lang.Object ref = hiveQueryId_; if (ref instanceof java.lang.String) { @@ -1053,7 +1261,9 @@ public java.lang.String getHiveQueryId() { } /** * optional string hiveQueryId = 2; + * @return The bytes for hiveQueryId. */ + @java.lang.Override public com.google.protobuf.ByteString getHiveQueryIdBytes() { java.lang.Object ref = hiveQueryId_; @@ -1068,34 +1278,40 @@ public java.lang.String getHiveQueryId() { } } - // optional int64 timestamp = 3; public static final int TIMESTAMP_FIELD_NUMBER = 3; private long timestamp_; /** * optional int64 timestamp = 3; + * @return Whether the timestamp field is set. */ + @java.lang.Override public boolean hasTimestamp() { - return ((bitField0_ & 0x00000004) == 0x00000004); + return ((bitField0_ & 0x00000004) != 0); } /** * optional int64 timestamp = 3; + * @return The timestamp. */ + @java.lang.Override public long getTimestamp() { return timestamp_; } - // optional string executionMode = 4; public static final int EXECUTIONMODE_FIELD_NUMBER = 4; - private java.lang.Object executionMode_; + private volatile java.lang.Object executionMode_; /** * optional string executionMode = 4; + * @return Whether the executionMode field is set. */ + @java.lang.Override public boolean hasExecutionMode() { - return ((bitField0_ & 0x00000008) == 0x00000008); + return ((bitField0_ & 0x00000008) != 0); } /** * optional string executionMode = 4; + * @return The executionMode. */ + @java.lang.Override public java.lang.String getExecutionMode() { java.lang.Object ref = executionMode_; if (ref instanceof java.lang.String) { @@ -1112,7 +1328,9 @@ public java.lang.String getExecutionMode() { } /** * optional string executionMode = 4; + * @return The bytes for executionMode. */ + @java.lang.Override public com.google.protobuf.ByteString getExecutionModeBytes() { java.lang.Object ref = executionMode_; @@ -1127,18 +1345,21 @@ public java.lang.String getExecutionMode() { } } - // optional string requestUser = 5; public static final int REQUESTUSER_FIELD_NUMBER = 5; - private java.lang.Object requestUser_; + private volatile java.lang.Object requestUser_; /** * optional string requestUser = 5; + * @return Whether the requestUser field is set. */ + @java.lang.Override public boolean hasRequestUser() { - return ((bitField0_ & 0x00000010) == 0x00000010); + return ((bitField0_ & 0x00000010) != 0); } /** * optional string requestUser = 5; + * @return The requestUser. */ + @java.lang.Override public java.lang.String getRequestUser() { java.lang.Object ref = requestUser_; if (ref instanceof java.lang.String) { @@ -1155,7 +1376,9 @@ public java.lang.String getRequestUser() { } /** * optional string requestUser = 5; + * @return The bytes for requestUser. */ + @java.lang.Override public com.google.protobuf.ByteString getRequestUserBytes() { java.lang.Object ref = requestUser_; @@ -1170,18 +1393,21 @@ public java.lang.String getRequestUser() { } } - // optional string queue = 6; public static final int QUEUE_FIELD_NUMBER = 6; - private java.lang.Object queue_; + private volatile java.lang.Object queue_; /** * optional string queue = 6; + * @return Whether the queue field is set. */ + @java.lang.Override public boolean hasQueue() { - return ((bitField0_ & 0x00000020) == 0x00000020); + return ((bitField0_ & 0x00000020) != 0); } /** * optional string queue = 6; + * @return The queue. */ + @java.lang.Override public java.lang.String getQueue() { java.lang.Object ref = queue_; if (ref instanceof java.lang.String) { @@ -1198,7 +1424,9 @@ public java.lang.String getQueue() { } /** * optional string queue = 6; + * @return The bytes for queue. */ + @java.lang.Override public com.google.protobuf.ByteString getQueueBytes() { java.lang.Object ref = queue_; @@ -1213,18 +1441,21 @@ public java.lang.String getQueue() { } } - // optional string user = 7; public static final int USER_FIELD_NUMBER = 7; - private java.lang.Object user_; + private volatile java.lang.Object user_; /** * optional string user = 7; + * @return Whether the user field is set. */ + @java.lang.Override public boolean hasUser() { - return ((bitField0_ & 0x00000040) == 0x00000040); + return ((bitField0_ & 0x00000040) != 0); } /** * optional string user = 7; + * @return The user. */ + @java.lang.Override public java.lang.String getUser() { java.lang.Object ref = user_; if (ref instanceof java.lang.String) { @@ -1241,7 +1472,9 @@ public java.lang.String getUser() { } /** * optional string user = 7; + * @return The bytes for user. */ + @java.lang.Override public com.google.protobuf.ByteString getUserBytes() { java.lang.Object ref = user_; @@ -1256,18 +1489,21 @@ public java.lang.String getUser() { } } - // optional string operationId = 8; public static final int OPERATIONID_FIELD_NUMBER = 8; - private java.lang.Object operationId_; + private volatile java.lang.Object operationId_; /** * optional string operationId = 8; + * @return Whether the operationId field is set. */ + @java.lang.Override public boolean hasOperationId() { - return ((bitField0_ & 0x00000080) == 0x00000080); + return ((bitField0_ & 0x00000080) != 0); } /** * optional string operationId = 8; + * @return The operationId. */ + @java.lang.Override public java.lang.String getOperationId() { java.lang.Object ref = operationId_; if (ref instanceof java.lang.String) { @@ -1284,7 +1520,9 @@ public java.lang.String getOperationId() { } /** * optional string operationId = 8; + * @return The bytes for operationId. */ + @java.lang.Override public com.google.protobuf.ByteString getOperationIdBytes() { java.lang.Object ref = operationId_; @@ -1299,206 +1537,201 @@ public java.lang.String getOperationId() { } } - // repeated string tablesWritten = 9; public static final int TABLESWRITTEN_FIELD_NUMBER = 9; private com.google.protobuf.LazyStringList tablesWritten_; /** * repeated string tablesWritten = 9; + * @return A list containing the tablesWritten. */ - public java.util.List + public com.google.protobuf.ProtocolStringList getTablesWrittenList() { return tablesWritten_; } /** * repeated string tablesWritten = 9; + * @return The count of tablesWritten. */ public int getTablesWrittenCount() { return tablesWritten_.size(); } /** * repeated string tablesWritten = 9; + * @param index The index of the element to return. + * @return The tablesWritten at the given index. */ public java.lang.String getTablesWritten(int index) { return tablesWritten_.get(index); } /** * repeated string tablesWritten = 9; + * @param index The index of the value to return. + * @return The bytes of the tablesWritten at the given index. */ public com.google.protobuf.ByteString getTablesWrittenBytes(int index) { return tablesWritten_.getByteString(index); } - // repeated string tablesRead = 10; public static final int TABLESREAD_FIELD_NUMBER = 10; private com.google.protobuf.LazyStringList tablesRead_; /** * repeated string tablesRead = 10; + * @return A list containing the tablesRead. */ - public java.util.List + public com.google.protobuf.ProtocolStringList getTablesReadList() { return tablesRead_; } /** * repeated string tablesRead = 10; + * @return The count of tablesRead. */ public int getTablesReadCount() { return tablesRead_.size(); } /** * repeated string tablesRead = 10; + * @param index The index of the element to return. + * @return The tablesRead at the given index. */ public java.lang.String getTablesRead(int index) { return tablesRead_.get(index); } /** * repeated string tablesRead = 10; + * @param index The index of the value to return. + * @return The bytes of the tablesRead at the given index. */ public com.google.protobuf.ByteString getTablesReadBytes(int index) { return tablesRead_.getByteString(index); } - // repeated .org.apache.hadoop.hive.ql.hooks.proto.MapFieldEntry otherInfo = 50; public static final int OTHERINFO_FIELD_NUMBER = 50; private java.util.List otherInfo_; /** - * repeated .org.apache.hadoop.hive.ql.hooks.proto.MapFieldEntry otherInfo = 50; + * repeated .MapFieldEntry otherInfo = 50; */ + @java.lang.Override public java.util.List getOtherInfoList() { return otherInfo_; } /** - * repeated .org.apache.hadoop.hive.ql.hooks.proto.MapFieldEntry otherInfo = 50; + * repeated .MapFieldEntry otherInfo = 50; */ + @java.lang.Override public java.util.List getOtherInfoOrBuilderList() { return otherInfo_; } /** - * repeated .org.apache.hadoop.hive.ql.hooks.proto.MapFieldEntry otherInfo = 50; + * repeated .MapFieldEntry otherInfo = 50; */ + @java.lang.Override public int getOtherInfoCount() { return otherInfo_.size(); } /** - * repeated .org.apache.hadoop.hive.ql.hooks.proto.MapFieldEntry otherInfo = 50; + * repeated .MapFieldEntry otherInfo = 50; */ + @java.lang.Override public org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry getOtherInfo(int index) { return otherInfo_.get(index); } /** - * repeated .org.apache.hadoop.hive.ql.hooks.proto.MapFieldEntry otherInfo = 50; + * repeated .MapFieldEntry otherInfo = 50; */ + @java.lang.Override public org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntryOrBuilder getOtherInfoOrBuilder( int index) { return otherInfo_.get(index); } - private void initFields() { - eventType_ = ""; - hiveQueryId_ = ""; - timestamp_ = 0L; - executionMode_ = ""; - requestUser_ = ""; - queue_ = ""; - user_ = ""; - operationId_ = ""; - tablesWritten_ = com.google.protobuf.LazyStringArrayList.EMPTY; - tablesRead_ = com.google.protobuf.LazyStringArrayList.EMPTY; - otherInfo_ = java.util.Collections.emptyList(); - } private byte memoizedIsInitialized = -1; + @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } + @java.lang.Override public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); - if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeBytes(1, getEventTypeBytes()); + if (((bitField0_ & 0x00000001) != 0)) { + com.google.protobuf.GeneratedMessageV3.writeString(output, 1, eventType_); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - output.writeBytes(2, getHiveQueryIdBytes()); + if (((bitField0_ & 0x00000002) != 0)) { + com.google.protobuf.GeneratedMessageV3.writeString(output, 2, hiveQueryId_); } - if (((bitField0_ & 0x00000004) == 0x00000004)) { + if (((bitField0_ & 0x00000004) != 0)) { output.writeInt64(3, timestamp_); } - if (((bitField0_ & 0x00000008) == 0x00000008)) { - output.writeBytes(4, getExecutionModeBytes()); + if (((bitField0_ & 0x00000008) != 0)) { + com.google.protobuf.GeneratedMessageV3.writeString(output, 4, executionMode_); } - if (((bitField0_ & 0x00000010) == 0x00000010)) { - output.writeBytes(5, getRequestUserBytes()); + if (((bitField0_ & 0x00000010) != 0)) { + com.google.protobuf.GeneratedMessageV3.writeString(output, 5, requestUser_); } - if (((bitField0_ & 0x00000020) == 0x00000020)) { - output.writeBytes(6, getQueueBytes()); + if (((bitField0_ & 0x00000020) != 0)) { + com.google.protobuf.GeneratedMessageV3.writeString(output, 6, queue_); } - if (((bitField0_ & 0x00000040) == 0x00000040)) { - output.writeBytes(7, getUserBytes()); + if (((bitField0_ & 0x00000040) != 0)) { + com.google.protobuf.GeneratedMessageV3.writeString(output, 7, user_); } - if (((bitField0_ & 0x00000080) == 0x00000080)) { - output.writeBytes(8, getOperationIdBytes()); + if (((bitField0_ & 0x00000080) != 0)) { + com.google.protobuf.GeneratedMessageV3.writeString(output, 8, operationId_); } for (int i = 0; i < tablesWritten_.size(); i++) { - output.writeBytes(9, tablesWritten_.getByteString(i)); + com.google.protobuf.GeneratedMessageV3.writeString(output, 9, tablesWritten_.getRaw(i)); } for (int i = 0; i < tablesRead_.size(); i++) { - output.writeBytes(10, tablesRead_.getByteString(i)); + com.google.protobuf.GeneratedMessageV3.writeString(output, 10, tablesRead_.getRaw(i)); } for (int i = 0; i < otherInfo_.size(); i++) { output.writeMessage(50, otherInfo_.get(i)); } - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; + @java.lang.Override public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(1, getEventTypeBytes()); + if (((bitField0_ & 0x00000001) != 0)) { + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, eventType_); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(2, getHiveQueryIdBytes()); + if (((bitField0_ & 0x00000002) != 0)) { + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(2, hiveQueryId_); } - if (((bitField0_ & 0x00000004) == 0x00000004)) { + if (((bitField0_ & 0x00000004) != 0)) { size += com.google.protobuf.CodedOutputStream .computeInt64Size(3, timestamp_); } - if (((bitField0_ & 0x00000008) == 0x00000008)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(4, getExecutionModeBytes()); + if (((bitField0_ & 0x00000008) != 0)) { + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(4, executionMode_); } - if (((bitField0_ & 0x00000010) == 0x00000010)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(5, getRequestUserBytes()); + if (((bitField0_ & 0x00000010) != 0)) { + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(5, requestUser_); } - if (((bitField0_ & 0x00000020) == 0x00000020)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(6, getQueueBytes()); + if (((bitField0_ & 0x00000020) != 0)) { + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(6, queue_); } - if (((bitField0_ & 0x00000040) == 0x00000040)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(7, getUserBytes()); + if (((bitField0_ & 0x00000040) != 0)) { + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(7, user_); } - if (((bitField0_ & 0x00000080) == 0x00000080)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(8, getOperationIdBytes()); + if (((bitField0_ & 0x00000080) != 0)) { + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(8, operationId_); } { int dataSize = 0; for (int i = 0; i < tablesWritten_.size(); i++) { - dataSize += com.google.protobuf.CodedOutputStream - .computeBytesSizeNoTag(tablesWritten_.getByteString(i)); + dataSize += computeStringSizeNoTag(tablesWritten_.getRaw(i)); } size += dataSize; size += 1 * getTablesWrittenList().size(); @@ -1506,8 +1739,7 @@ public int getSerializedSize() { { int dataSize = 0; for (int i = 0; i < tablesRead_.size(); i++) { - dataSize += com.google.protobuf.CodedOutputStream - .computeBytesSizeNoTag(tablesRead_.getByteString(i)); + dataSize += computeStringSizeNoTag(tablesRead_.getRaw(i)); } size += dataSize; size += 1 * getTablesReadList().size(); @@ -1516,18 +1748,139 @@ public int getSerializedSize() { size += com.google.protobuf.CodedOutputStream .computeMessageSize(50, otherInfo_.get(i)); } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } - private static final long serialVersionUID = 0L; @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); + public boolean equals(final java.lang.Object obj) { + if (obj == this) { + return true; + } + if (!(obj instanceof org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.HiveHookEventProto)) { + return super.equals(obj); + } + org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.HiveHookEventProto other = (org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.HiveHookEventProto) obj; + + if (hasEventType() != other.hasEventType()) return false; + if (hasEventType()) { + if (!getEventType() + .equals(other.getEventType())) return false; + } + if (hasHiveQueryId() != other.hasHiveQueryId()) return false; + if (hasHiveQueryId()) { + if (!getHiveQueryId() + .equals(other.getHiveQueryId())) return false; + } + if (hasTimestamp() != other.hasTimestamp()) return false; + if (hasTimestamp()) { + if (getTimestamp() + != other.getTimestamp()) return false; + } + if (hasExecutionMode() != other.hasExecutionMode()) return false; + if (hasExecutionMode()) { + if (!getExecutionMode() + .equals(other.getExecutionMode())) return false; + } + if (hasRequestUser() != other.hasRequestUser()) return false; + if (hasRequestUser()) { + if (!getRequestUser() + .equals(other.getRequestUser())) return false; + } + if (hasQueue() != other.hasQueue()) return false; + if (hasQueue()) { + if (!getQueue() + .equals(other.getQueue())) return false; + } + if (hasUser() != other.hasUser()) return false; + if (hasUser()) { + if (!getUser() + .equals(other.getUser())) return false; + } + if (hasOperationId() != other.hasOperationId()) return false; + if (hasOperationId()) { + if (!getOperationId() + .equals(other.getOperationId())) return false; + } + if (!getTablesWrittenList() + .equals(other.getTablesWrittenList())) return false; + if (!getTablesReadList() + .equals(other.getTablesReadList())) return false; + if (!getOtherInfoList() + .equals(other.getOtherInfoList())) return false; + if (!unknownFields.equals(other.unknownFields)) return false; + return true; + } + + @java.lang.Override + public int hashCode() { + if (memoizedHashCode != 0) { + return memoizedHashCode; + } + int hash = 41; + hash = (19 * hash) + getDescriptor().hashCode(); + if (hasEventType()) { + hash = (37 * hash) + EVENTTYPE_FIELD_NUMBER; + hash = (53 * hash) + getEventType().hashCode(); + } + if (hasHiveQueryId()) { + hash = (37 * hash) + HIVEQUERYID_FIELD_NUMBER; + hash = (53 * hash) + getHiveQueryId().hashCode(); + } + if (hasTimestamp()) { + hash = (37 * hash) + TIMESTAMP_FIELD_NUMBER; + hash = (53 * hash) + com.google.protobuf.Internal.hashLong( + getTimestamp()); + } + if (hasExecutionMode()) { + hash = (37 * hash) + EXECUTIONMODE_FIELD_NUMBER; + hash = (53 * hash) + getExecutionMode().hashCode(); + } + if (hasRequestUser()) { + hash = (37 * hash) + REQUESTUSER_FIELD_NUMBER; + hash = (53 * hash) + getRequestUser().hashCode(); + } + if (hasQueue()) { + hash = (37 * hash) + QUEUE_FIELD_NUMBER; + hash = (53 * hash) + getQueue().hashCode(); + } + if (hasUser()) { + hash = (37 * hash) + USER_FIELD_NUMBER; + hash = (53 * hash) + getUser().hashCode(); + } + if (hasOperationId()) { + hash = (37 * hash) + OPERATIONID_FIELD_NUMBER; + hash = (53 * hash) + getOperationId().hashCode(); + } + if (getTablesWrittenCount() > 0) { + hash = (37 * hash) + TABLESWRITTEN_FIELD_NUMBER; + hash = (53 * hash) + getTablesWrittenList().hashCode(); + } + if (getTablesReadCount() > 0) { + hash = (37 * hash) + TABLESREAD_FIELD_NUMBER; + hash = (53 * hash) + getTablesReadList().hashCode(); + } + if (getOtherInfoCount() > 0) { + hash = (37 * hash) + OTHERINFO_FIELD_NUMBER; + hash = (53 * hash) + getOtherInfoList().hashCode(); + } + hash = (29 * hash) + unknownFields.hashCode(); + memoizedHashCode = hash; + return hash; } + public static org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.HiveHookEventProto parseFrom( + java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.HiveHookEventProto parseFrom( + java.nio.ByteBuffer data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } public static org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.HiveHookEventProto parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { @@ -1551,63 +1904,78 @@ public static org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.HiveHookEvent } public static org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.HiveHookEventProto parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.HiveHookEventProto parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.HiveHookEventProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.HiveHookEventProto parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.HiveHookEventProto parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.HiveHookEventProto parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } + @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.HiveHookEventProto prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** - * Protobuf type {@code org.apache.hadoop.hive.ql.hooks.proto.HiveHookEventProto} + * Protobuf type {@code HiveHookEventProto} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.HiveHookEventProtoOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder implements + // @@protoc_insertion_point(builder_implements:HiveHookEventProto) + org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.HiveHookEventProtoOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { - return org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.internal_static_org_apache_hadoop_hive_ql_hooks_proto_HiveHookEventProto_descriptor; + return org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.internal_static_HiveHookEventProto_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { - return org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.internal_static_org_apache_hadoop_hive_ql_hooks_proto_HiveHookEventProto_fieldAccessorTable + return org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.internal_static_HiveHookEventProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.HiveHookEventProto.class, org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.HiveHookEventProto.Builder.class); } @@ -1618,19 +1986,17 @@ private Builder() { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { getOtherInfoFieldBuilder(); } } - private static Builder create() { - return new Builder(); - } - + @java.lang.Override public Builder clear() { super.clear(); eventType_ = ""; @@ -1662,19 +2028,18 @@ public Builder clear() { return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - + @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { - return org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.internal_static_org_apache_hadoop_hive_ql_hooks_proto_HiveHookEventProto_descriptor; + return org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.internal_static_HiveHookEventProto_descriptor; } + @java.lang.Override public org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.HiveHookEventProto getDefaultInstanceForType() { return org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.HiveHookEventProto.getDefaultInstance(); } + @java.lang.Override public org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.HiveHookEventProto build() { org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.HiveHookEventProto result = buildPartial(); if (!result.isInitialized()) { @@ -1683,56 +2048,55 @@ public org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.HiveHookEventProto b return result; } + @java.lang.Override public org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.HiveHookEventProto buildPartial() { org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.HiveHookEventProto result = new org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.HiveHookEventProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { + if (((from_bitField0_ & 0x00000001) != 0)) { to_bitField0_ |= 0x00000001; } result.eventType_ = eventType_; - if (((from_bitField0_ & 0x00000002) == 0x00000002)) { + if (((from_bitField0_ & 0x00000002) != 0)) { to_bitField0_ |= 0x00000002; } result.hiveQueryId_ = hiveQueryId_; - if (((from_bitField0_ & 0x00000004) == 0x00000004)) { + if (((from_bitField0_ & 0x00000004) != 0)) { + result.timestamp_ = timestamp_; to_bitField0_ |= 0x00000004; } - result.timestamp_ = timestamp_; - if (((from_bitField0_ & 0x00000008) == 0x00000008)) { + if (((from_bitField0_ & 0x00000008) != 0)) { to_bitField0_ |= 0x00000008; } result.executionMode_ = executionMode_; - if (((from_bitField0_ & 0x00000010) == 0x00000010)) { + if (((from_bitField0_ & 0x00000010) != 0)) { to_bitField0_ |= 0x00000010; } result.requestUser_ = requestUser_; - if (((from_bitField0_ & 0x00000020) == 0x00000020)) { + if (((from_bitField0_ & 0x00000020) != 0)) { to_bitField0_ |= 0x00000020; } result.queue_ = queue_; - if (((from_bitField0_ & 0x00000040) == 0x00000040)) { + if (((from_bitField0_ & 0x00000040) != 0)) { to_bitField0_ |= 0x00000040; } result.user_ = user_; - if (((from_bitField0_ & 0x00000080) == 0x00000080)) { + if (((from_bitField0_ & 0x00000080) != 0)) { to_bitField0_ |= 0x00000080; } result.operationId_ = operationId_; - if (((bitField0_ & 0x00000100) == 0x00000100)) { - tablesWritten_ = new com.google.protobuf.UnmodifiableLazyStringList( - tablesWritten_); + if (((bitField0_ & 0x00000100) != 0)) { + tablesWritten_ = tablesWritten_.getUnmodifiableView(); bitField0_ = (bitField0_ & ~0x00000100); } result.tablesWritten_ = tablesWritten_; - if (((bitField0_ & 0x00000200) == 0x00000200)) { - tablesRead_ = new com.google.protobuf.UnmodifiableLazyStringList( - tablesRead_); + if (((bitField0_ & 0x00000200) != 0)) { + tablesRead_ = tablesRead_.getUnmodifiableView(); bitField0_ = (bitField0_ & ~0x00000200); } result.tablesRead_ = tablesRead_; if (otherInfoBuilder_ == null) { - if (((bitField0_ & 0x00000400) == 0x00000400)) { + if (((bitField0_ & 0x00000400) != 0)) { otherInfo_ = java.util.Collections.unmodifiableList(otherInfo_); bitField0_ = (bitField0_ & ~0x00000400); } @@ -1745,6 +2109,39 @@ public org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.HiveHookEventProto b return result; } + @java.lang.Override + public Builder clone() { + return super.clone(); + } + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.setField(field, value); + } + @java.lang.Override + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + @java.lang.Override + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.addRepeatedField(field, value); + } + @java.lang.Override public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.HiveHookEventProto) { return mergeFrom((org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.HiveHookEventProto)other); @@ -1833,21 +2230,24 @@ public Builder mergeFrom(org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.Hi otherInfo_ = other.otherInfo_; bitField0_ = (bitField0_ & ~0x00000400); otherInfoBuilder_ = - com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? + com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getOtherInfoFieldBuilder() : null; } else { otherInfoBuilder_.addAllMessages(other.otherInfo_); } } } - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } + @java.lang.Override public final boolean isInitialized() { return true; } + @java.lang.Override public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) @@ -1857,7 +2257,7 @@ public Builder mergeFrom( parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.HiveHookEventProto) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -1867,23 +2267,27 @@ public Builder mergeFrom( } private int bitField0_; - // optional string eventType = 1; private java.lang.Object eventType_ = ""; /** * optional string eventType = 1; + * @return Whether the eventType field is set. */ public boolean hasEventType() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional string eventType = 1; + * @return The eventType. */ public java.lang.String getEventType() { java.lang.Object ref = eventType_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - eventType_ = s; + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + eventType_ = s; + } return s; } else { return (java.lang.String) ref; @@ -1891,6 +2295,7 @@ public java.lang.String getEventType() { } /** * optional string eventType = 1; + * @return The bytes for eventType. */ public com.google.protobuf.ByteString getEventTypeBytes() { @@ -1907,6 +2312,8 @@ public java.lang.String getEventType() { } /** * optional string eventType = 1; + * @param value The eventType to set. + * @return This builder for chaining. */ public Builder setEventType( java.lang.String value) { @@ -1920,6 +2327,7 @@ public Builder setEventType( } /** * optional string eventType = 1; + * @return This builder for chaining. */ public Builder clearEventType() { bitField0_ = (bitField0_ & ~0x00000001); @@ -1929,6 +2337,8 @@ public Builder clearEventType() { } /** * optional string eventType = 1; + * @param value The bytes for eventType to set. + * @return This builder for chaining. */ public Builder setEventTypeBytes( com.google.protobuf.ByteString value) { @@ -1941,23 +2351,27 @@ public Builder setEventTypeBytes( return this; } - // optional string hiveQueryId = 2; private java.lang.Object hiveQueryId_ = ""; /** * optional string hiveQueryId = 2; + * @return Whether the hiveQueryId field is set. */ public boolean hasHiveQueryId() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * optional string hiveQueryId = 2; + * @return The hiveQueryId. */ public java.lang.String getHiveQueryId() { java.lang.Object ref = hiveQueryId_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - hiveQueryId_ = s; + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + hiveQueryId_ = s; + } return s; } else { return (java.lang.String) ref; @@ -1965,6 +2379,7 @@ public java.lang.String getHiveQueryId() { } /** * optional string hiveQueryId = 2; + * @return The bytes for hiveQueryId. */ public com.google.protobuf.ByteString getHiveQueryIdBytes() { @@ -1981,6 +2396,8 @@ public java.lang.String getHiveQueryId() { } /** * optional string hiveQueryId = 2; + * @param value The hiveQueryId to set. + * @return This builder for chaining. */ public Builder setHiveQueryId( java.lang.String value) { @@ -1994,6 +2411,7 @@ public Builder setHiveQueryId( } /** * optional string hiveQueryId = 2; + * @return This builder for chaining. */ public Builder clearHiveQueryId() { bitField0_ = (bitField0_ & ~0x00000002); @@ -2003,6 +2421,8 @@ public Builder clearHiveQueryId() { } /** * optional string hiveQueryId = 2; + * @param value The bytes for hiveQueryId to set. + * @return This builder for chaining. */ public Builder setHiveQueryIdBytes( com.google.protobuf.ByteString value) { @@ -2015,22 +2435,27 @@ public Builder setHiveQueryIdBytes( return this; } - // optional int64 timestamp = 3; private long timestamp_ ; /** * optional int64 timestamp = 3; + * @return Whether the timestamp field is set. */ + @java.lang.Override public boolean hasTimestamp() { - return ((bitField0_ & 0x00000004) == 0x00000004); + return ((bitField0_ & 0x00000004) != 0); } /** * optional int64 timestamp = 3; + * @return The timestamp. */ + @java.lang.Override public long getTimestamp() { return timestamp_; } /** * optional int64 timestamp = 3; + * @param value The timestamp to set. + * @return This builder for chaining. */ public Builder setTimestamp(long value) { bitField0_ |= 0x00000004; @@ -2040,6 +2465,7 @@ public Builder setTimestamp(long value) { } /** * optional int64 timestamp = 3; + * @return This builder for chaining. */ public Builder clearTimestamp() { bitField0_ = (bitField0_ & ~0x00000004); @@ -2048,23 +2474,27 @@ public Builder clearTimestamp() { return this; } - // optional string executionMode = 4; private java.lang.Object executionMode_ = ""; /** * optional string executionMode = 4; + * @return Whether the executionMode field is set. */ public boolean hasExecutionMode() { - return ((bitField0_ & 0x00000008) == 0x00000008); + return ((bitField0_ & 0x00000008) != 0); } /** * optional string executionMode = 4; + * @return The executionMode. */ public java.lang.String getExecutionMode() { java.lang.Object ref = executionMode_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - executionMode_ = s; + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + executionMode_ = s; + } return s; } else { return (java.lang.String) ref; @@ -2072,6 +2502,7 @@ public java.lang.String getExecutionMode() { } /** * optional string executionMode = 4; + * @return The bytes for executionMode. */ public com.google.protobuf.ByteString getExecutionModeBytes() { @@ -2088,6 +2519,8 @@ public java.lang.String getExecutionMode() { } /** * optional string executionMode = 4; + * @param value The executionMode to set. + * @return This builder for chaining. */ public Builder setExecutionMode( java.lang.String value) { @@ -2101,6 +2534,7 @@ public Builder setExecutionMode( } /** * optional string executionMode = 4; + * @return This builder for chaining. */ public Builder clearExecutionMode() { bitField0_ = (bitField0_ & ~0x00000008); @@ -2110,6 +2544,8 @@ public Builder clearExecutionMode() { } /** * optional string executionMode = 4; + * @param value The bytes for executionMode to set. + * @return This builder for chaining. */ public Builder setExecutionModeBytes( com.google.protobuf.ByteString value) { @@ -2122,23 +2558,27 @@ public Builder setExecutionModeBytes( return this; } - // optional string requestUser = 5; private java.lang.Object requestUser_ = ""; /** * optional string requestUser = 5; + * @return Whether the requestUser field is set. */ public boolean hasRequestUser() { - return ((bitField0_ & 0x00000010) == 0x00000010); + return ((bitField0_ & 0x00000010) != 0); } /** * optional string requestUser = 5; + * @return The requestUser. */ public java.lang.String getRequestUser() { java.lang.Object ref = requestUser_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - requestUser_ = s; + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + requestUser_ = s; + } return s; } else { return (java.lang.String) ref; @@ -2146,6 +2586,7 @@ public java.lang.String getRequestUser() { } /** * optional string requestUser = 5; + * @return The bytes for requestUser. */ public com.google.protobuf.ByteString getRequestUserBytes() { @@ -2162,6 +2603,8 @@ public java.lang.String getRequestUser() { } /** * optional string requestUser = 5; + * @param value The requestUser to set. + * @return This builder for chaining. */ public Builder setRequestUser( java.lang.String value) { @@ -2175,6 +2618,7 @@ public Builder setRequestUser( } /** * optional string requestUser = 5; + * @return This builder for chaining. */ public Builder clearRequestUser() { bitField0_ = (bitField0_ & ~0x00000010); @@ -2184,6 +2628,8 @@ public Builder clearRequestUser() { } /** * optional string requestUser = 5; + * @param value The bytes for requestUser to set. + * @return This builder for chaining. */ public Builder setRequestUserBytes( com.google.protobuf.ByteString value) { @@ -2196,23 +2642,27 @@ public Builder setRequestUserBytes( return this; } - // optional string queue = 6; private java.lang.Object queue_ = ""; /** * optional string queue = 6; + * @return Whether the queue field is set. */ public boolean hasQueue() { - return ((bitField0_ & 0x00000020) == 0x00000020); + return ((bitField0_ & 0x00000020) != 0); } /** * optional string queue = 6; + * @return The queue. */ public java.lang.String getQueue() { java.lang.Object ref = queue_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - queue_ = s; + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + queue_ = s; + } return s; } else { return (java.lang.String) ref; @@ -2220,6 +2670,7 @@ public java.lang.String getQueue() { } /** * optional string queue = 6; + * @return The bytes for queue. */ public com.google.protobuf.ByteString getQueueBytes() { @@ -2236,6 +2687,8 @@ public java.lang.String getQueue() { } /** * optional string queue = 6; + * @param value The queue to set. + * @return This builder for chaining. */ public Builder setQueue( java.lang.String value) { @@ -2249,6 +2702,7 @@ public Builder setQueue( } /** * optional string queue = 6; + * @return This builder for chaining. */ public Builder clearQueue() { bitField0_ = (bitField0_ & ~0x00000020); @@ -2258,6 +2712,8 @@ public Builder clearQueue() { } /** * optional string queue = 6; + * @param value The bytes for queue to set. + * @return This builder for chaining. */ public Builder setQueueBytes( com.google.protobuf.ByteString value) { @@ -2270,23 +2726,27 @@ public Builder setQueueBytes( return this; } - // optional string user = 7; private java.lang.Object user_ = ""; /** * optional string user = 7; + * @return Whether the user field is set. */ public boolean hasUser() { - return ((bitField0_ & 0x00000040) == 0x00000040); + return ((bitField0_ & 0x00000040) != 0); } /** * optional string user = 7; + * @return The user. */ public java.lang.String getUser() { java.lang.Object ref = user_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - user_ = s; + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + user_ = s; + } return s; } else { return (java.lang.String) ref; @@ -2294,6 +2754,7 @@ public java.lang.String getUser() { } /** * optional string user = 7; + * @return The bytes for user. */ public com.google.protobuf.ByteString getUserBytes() { @@ -2310,6 +2771,8 @@ public java.lang.String getUser() { } /** * optional string user = 7; + * @param value The user to set. + * @return This builder for chaining. */ public Builder setUser( java.lang.String value) { @@ -2323,6 +2786,7 @@ public Builder setUser( } /** * optional string user = 7; + * @return This builder for chaining. */ public Builder clearUser() { bitField0_ = (bitField0_ & ~0x00000040); @@ -2332,6 +2796,8 @@ public Builder clearUser() { } /** * optional string user = 7; + * @param value The bytes for user to set. + * @return This builder for chaining. */ public Builder setUserBytes( com.google.protobuf.ByteString value) { @@ -2344,23 +2810,27 @@ public Builder setUserBytes( return this; } - // optional string operationId = 8; private java.lang.Object operationId_ = ""; /** * optional string operationId = 8; + * @return Whether the operationId field is set. */ public boolean hasOperationId() { - return ((bitField0_ & 0x00000080) == 0x00000080); + return ((bitField0_ & 0x00000080) != 0); } /** * optional string operationId = 8; + * @return The operationId. */ public java.lang.String getOperationId() { java.lang.Object ref = operationId_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - operationId_ = s; + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + operationId_ = s; + } return s; } else { return (java.lang.String) ref; @@ -2368,6 +2838,7 @@ public java.lang.String getOperationId() { } /** * optional string operationId = 8; + * @return The bytes for operationId. */ public com.google.protobuf.ByteString getOperationIdBytes() { @@ -2384,6 +2855,8 @@ public java.lang.String getOperationId() { } /** * optional string operationId = 8; + * @param value The operationId to set. + * @return This builder for chaining. */ public Builder setOperationId( java.lang.String value) { @@ -2397,6 +2870,7 @@ public Builder setOperationId( } /** * optional string operationId = 8; + * @return This builder for chaining. */ public Builder clearOperationId() { bitField0_ = (bitField0_ & ~0x00000080); @@ -2406,6 +2880,8 @@ public Builder clearOperationId() { } /** * optional string operationId = 8; + * @param value The bytes for operationId to set. + * @return This builder for chaining. */ public Builder setOperationIdBytes( com.google.protobuf.ByteString value) { @@ -2418,35 +2894,40 @@ public Builder setOperationIdBytes( return this; } - // repeated string tablesWritten = 9; private com.google.protobuf.LazyStringList tablesWritten_ = com.google.protobuf.LazyStringArrayList.EMPTY; private void ensureTablesWrittenIsMutable() { - if (!((bitField0_ & 0x00000100) == 0x00000100)) { + if (!((bitField0_ & 0x00000100) != 0)) { tablesWritten_ = new com.google.protobuf.LazyStringArrayList(tablesWritten_); bitField0_ |= 0x00000100; } } /** * repeated string tablesWritten = 9; + * @return A list containing the tablesWritten. */ - public java.util.List + public com.google.protobuf.ProtocolStringList getTablesWrittenList() { - return java.util.Collections.unmodifiableList(tablesWritten_); + return tablesWritten_.getUnmodifiableView(); } /** * repeated string tablesWritten = 9; + * @return The count of tablesWritten. */ public int getTablesWrittenCount() { return tablesWritten_.size(); } /** * repeated string tablesWritten = 9; + * @param index The index of the element to return. + * @return The tablesWritten at the given index. */ public java.lang.String getTablesWritten(int index) { return tablesWritten_.get(index); } /** * repeated string tablesWritten = 9; + * @param index The index of the value to return. + * @return The bytes of the tablesWritten at the given index. */ public com.google.protobuf.ByteString getTablesWrittenBytes(int index) { @@ -2454,6 +2935,9 @@ public java.lang.String getTablesWritten(int index) { } /** * repeated string tablesWritten = 9; + * @param index The index to set the value at. + * @param value The tablesWritten to set. + * @return This builder for chaining. */ public Builder setTablesWritten( int index, java.lang.String value) { @@ -2467,6 +2951,8 @@ public Builder setTablesWritten( } /** * repeated string tablesWritten = 9; + * @param value The tablesWritten to add. + * @return This builder for chaining. */ public Builder addTablesWritten( java.lang.String value) { @@ -2480,16 +2966,20 @@ public Builder addTablesWritten( } /** * repeated string tablesWritten = 9; + * @param values The tablesWritten to add. + * @return This builder for chaining. */ public Builder addAllTablesWritten( java.lang.Iterable values) { ensureTablesWrittenIsMutable(); - super.addAll(values, tablesWritten_); + com.google.protobuf.AbstractMessageLite.Builder.addAll( + values, tablesWritten_); onChanged(); return this; } /** * repeated string tablesWritten = 9; + * @return This builder for chaining. */ public Builder clearTablesWritten() { tablesWritten_ = com.google.protobuf.LazyStringArrayList.EMPTY; @@ -2499,6 +2989,8 @@ public Builder clearTablesWritten() { } /** * repeated string tablesWritten = 9; + * @param value The bytes of the tablesWritten to add. + * @return This builder for chaining. */ public Builder addTablesWrittenBytes( com.google.protobuf.ByteString value) { @@ -2511,35 +3003,40 @@ public Builder addTablesWrittenBytes( return this; } - // repeated string tablesRead = 10; private com.google.protobuf.LazyStringList tablesRead_ = com.google.protobuf.LazyStringArrayList.EMPTY; private void ensureTablesReadIsMutable() { - if (!((bitField0_ & 0x00000200) == 0x00000200)) { + if (!((bitField0_ & 0x00000200) != 0)) { tablesRead_ = new com.google.protobuf.LazyStringArrayList(tablesRead_); bitField0_ |= 0x00000200; } } /** * repeated string tablesRead = 10; + * @return A list containing the tablesRead. */ - public java.util.List + public com.google.protobuf.ProtocolStringList getTablesReadList() { - return java.util.Collections.unmodifiableList(tablesRead_); + return tablesRead_.getUnmodifiableView(); } /** * repeated string tablesRead = 10; + * @return The count of tablesRead. */ public int getTablesReadCount() { return tablesRead_.size(); } /** * repeated string tablesRead = 10; + * @param index The index of the element to return. + * @return The tablesRead at the given index. */ public java.lang.String getTablesRead(int index) { return tablesRead_.get(index); } /** * repeated string tablesRead = 10; + * @param index The index of the value to return. + * @return The bytes of the tablesRead at the given index. */ public com.google.protobuf.ByteString getTablesReadBytes(int index) { @@ -2547,6 +3044,9 @@ public java.lang.String getTablesRead(int index) { } /** * repeated string tablesRead = 10; + * @param index The index to set the value at. + * @param value The tablesRead to set. + * @return This builder for chaining. */ public Builder setTablesRead( int index, java.lang.String value) { @@ -2560,6 +3060,8 @@ public Builder setTablesRead( } /** * repeated string tablesRead = 10; + * @param value The tablesRead to add. + * @return This builder for chaining. */ public Builder addTablesRead( java.lang.String value) { @@ -2573,16 +3075,20 @@ public Builder addTablesRead( } /** * repeated string tablesRead = 10; + * @param values The tablesRead to add. + * @return This builder for chaining. */ public Builder addAllTablesRead( java.lang.Iterable values) { ensureTablesReadIsMutable(); - super.addAll(values, tablesRead_); + com.google.protobuf.AbstractMessageLite.Builder.addAll( + values, tablesRead_); onChanged(); return this; } /** * repeated string tablesRead = 10; + * @return This builder for chaining. */ public Builder clearTablesRead() { tablesRead_ = com.google.protobuf.LazyStringArrayList.EMPTY; @@ -2592,6 +3098,8 @@ public Builder clearTablesRead() { } /** * repeated string tablesRead = 10; + * @param value The bytes of the tablesRead to add. + * @return This builder for chaining. */ public Builder addTablesReadBytes( com.google.protobuf.ByteString value) { @@ -2604,21 +3112,20 @@ public Builder addTablesReadBytes( return this; } - // repeated .org.apache.hadoop.hive.ql.hooks.proto.MapFieldEntry otherInfo = 50; private java.util.List otherInfo_ = java.util.Collections.emptyList(); private void ensureOtherInfoIsMutable() { - if (!((bitField0_ & 0x00000400) == 0x00000400)) { + if (!((bitField0_ & 0x00000400) != 0)) { otherInfo_ = new java.util.ArrayList(otherInfo_); bitField0_ |= 0x00000400; } } - private com.google.protobuf.RepeatedFieldBuilder< + private com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry, org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry.Builder, org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntryOrBuilder> otherInfoBuilder_; /** - * repeated .org.apache.hadoop.hive.ql.hooks.proto.MapFieldEntry otherInfo = 50; + * repeated .MapFieldEntry otherInfo = 50; */ public java.util.List getOtherInfoList() { if (otherInfoBuilder_ == null) { @@ -2628,7 +3135,7 @@ public java.util.Listrepeated .org.apache.hadoop.hive.ql.hooks.proto.MapFieldEntry otherInfo = 50; + * repeated .MapFieldEntry otherInfo = 50; */ public int getOtherInfoCount() { if (otherInfoBuilder_ == null) { @@ -2638,7 +3145,7 @@ public int getOtherInfoCount() { } } /** - * repeated .org.apache.hadoop.hive.ql.hooks.proto.MapFieldEntry otherInfo = 50; + * repeated .MapFieldEntry otherInfo = 50; */ public org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry getOtherInfo(int index) { if (otherInfoBuilder_ == null) { @@ -2648,7 +3155,7 @@ public org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry getOth } } /** - * repeated .org.apache.hadoop.hive.ql.hooks.proto.MapFieldEntry otherInfo = 50; + * repeated .MapFieldEntry otherInfo = 50; */ public Builder setOtherInfo( int index, org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry value) { @@ -2665,7 +3172,7 @@ public Builder setOtherInfo( return this; } /** - * repeated .org.apache.hadoop.hive.ql.hooks.proto.MapFieldEntry otherInfo = 50; + * repeated .MapFieldEntry otherInfo = 50; */ public Builder setOtherInfo( int index, org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry.Builder builderForValue) { @@ -2679,7 +3186,7 @@ public Builder setOtherInfo( return this; } /** - * repeated .org.apache.hadoop.hive.ql.hooks.proto.MapFieldEntry otherInfo = 50; + * repeated .MapFieldEntry otherInfo = 50; */ public Builder addOtherInfo(org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry value) { if (otherInfoBuilder_ == null) { @@ -2695,7 +3202,7 @@ public Builder addOtherInfo(org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents return this; } /** - * repeated .org.apache.hadoop.hive.ql.hooks.proto.MapFieldEntry otherInfo = 50; + * repeated .MapFieldEntry otherInfo = 50; */ public Builder addOtherInfo( int index, org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry value) { @@ -2712,7 +3219,7 @@ public Builder addOtherInfo( return this; } /** - * repeated .org.apache.hadoop.hive.ql.hooks.proto.MapFieldEntry otherInfo = 50; + * repeated .MapFieldEntry otherInfo = 50; */ public Builder addOtherInfo( org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry.Builder builderForValue) { @@ -2726,7 +3233,7 @@ public Builder addOtherInfo( return this; } /** - * repeated .org.apache.hadoop.hive.ql.hooks.proto.MapFieldEntry otherInfo = 50; + * repeated .MapFieldEntry otherInfo = 50; */ public Builder addOtherInfo( int index, org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry.Builder builderForValue) { @@ -2740,13 +3247,14 @@ public Builder addOtherInfo( return this; } /** - * repeated .org.apache.hadoop.hive.ql.hooks.proto.MapFieldEntry otherInfo = 50; + * repeated .MapFieldEntry otherInfo = 50; */ public Builder addAllOtherInfo( java.lang.Iterable values) { if (otherInfoBuilder_ == null) { ensureOtherInfoIsMutable(); - super.addAll(values, otherInfo_); + com.google.protobuf.AbstractMessageLite.Builder.addAll( + values, otherInfo_); onChanged(); } else { otherInfoBuilder_.addAllMessages(values); @@ -2754,7 +3262,7 @@ public Builder addAllOtherInfo( return this; } /** - * repeated .org.apache.hadoop.hive.ql.hooks.proto.MapFieldEntry otherInfo = 50; + * repeated .MapFieldEntry otherInfo = 50; */ public Builder clearOtherInfo() { if (otherInfoBuilder_ == null) { @@ -2767,7 +3275,7 @@ public Builder clearOtherInfo() { return this; } /** - * repeated .org.apache.hadoop.hive.ql.hooks.proto.MapFieldEntry otherInfo = 50; + * repeated .MapFieldEntry otherInfo = 50; */ public Builder removeOtherInfo(int index) { if (otherInfoBuilder_ == null) { @@ -2780,14 +3288,14 @@ public Builder removeOtherInfo(int index) { return this; } /** - * repeated .org.apache.hadoop.hive.ql.hooks.proto.MapFieldEntry otherInfo = 50; + * repeated .MapFieldEntry otherInfo = 50; */ public org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry.Builder getOtherInfoBuilder( int index) { return getOtherInfoFieldBuilder().getBuilder(index); } /** - * repeated .org.apache.hadoop.hive.ql.hooks.proto.MapFieldEntry otherInfo = 50; + * repeated .MapFieldEntry otherInfo = 50; */ public org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntryOrBuilder getOtherInfoOrBuilder( int index) { @@ -2797,7 +3305,7 @@ public org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntryOrBuild } } /** - * repeated .org.apache.hadoop.hive.ql.hooks.proto.MapFieldEntry otherInfo = 50; + * repeated .MapFieldEntry otherInfo = 50; */ public java.util.List getOtherInfoOrBuilderList() { @@ -2808,14 +3316,14 @@ public org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntryOrBuild } } /** - * repeated .org.apache.hadoop.hive.ql.hooks.proto.MapFieldEntry otherInfo = 50; + * repeated .MapFieldEntry otherInfo = 50; */ public org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry.Builder addOtherInfoBuilder() { return getOtherInfoFieldBuilder().addBuilder( org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry.getDefaultInstance()); } /** - * repeated .org.apache.hadoop.hive.ql.hooks.proto.MapFieldEntry otherInfo = 50; + * repeated .MapFieldEntry otherInfo = 50; */ public org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry.Builder addOtherInfoBuilder( int index) { @@ -2823,94 +3331,125 @@ public org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry.Builde index, org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry.getDefaultInstance()); } /** - * repeated .org.apache.hadoop.hive.ql.hooks.proto.MapFieldEntry otherInfo = 50; + * repeated .MapFieldEntry otherInfo = 50; */ public java.util.List getOtherInfoBuilderList() { return getOtherInfoFieldBuilder().getBuilderList(); } - private com.google.protobuf.RepeatedFieldBuilder< + private com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry, org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry.Builder, org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntryOrBuilder> getOtherInfoFieldBuilder() { if (otherInfoBuilder_ == null) { - otherInfoBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< + otherInfoBuilder_ = new com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry, org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntry.Builder, org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.MapFieldEntryOrBuilder>( otherInfo_, - ((bitField0_ & 0x00000400) == 0x00000400), + ((bitField0_ & 0x00000400) != 0), getParentForChildren(), isClean()); otherInfo_ = null; } return otherInfoBuilder_; } + @java.lang.Override + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } - // @@protoc_insertion_point(builder_scope:org.apache.hadoop.hive.ql.hooks.proto.HiveHookEventProto) + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + + + // @@protoc_insertion_point(builder_scope:HiveHookEventProto) } + // @@protoc_insertion_point(class_scope:HiveHookEventProto) + private static final org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.HiveHookEventProto DEFAULT_INSTANCE; static { - defaultInstance = new HiveHookEventProto(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.HiveHookEventProto(); + } + + public static org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.HiveHookEventProto getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser + PARSER = new com.google.protobuf.AbstractParser() { + @java.lang.Override + public HiveHookEventProto parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new HiveHookEventProto(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public org.apache.hadoop.hive.ql.hooks.proto.HiveHookEvents.HiveHookEventProto getDefaultInstanceForType() { + return DEFAULT_INSTANCE; } - // @@protoc_insertion_point(class_scope:org.apache.hadoop.hive.ql.hooks.proto.HiveHookEventProto) } - private static com.google.protobuf.Descriptors.Descriptor - internal_static_org_apache_hadoop_hive_ql_hooks_proto_MapFieldEntry_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable - internal_static_org_apache_hadoop_hive_ql_hooks_proto_MapFieldEntry_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor - internal_static_org_apache_hadoop_hive_ql_hooks_proto_HiveHookEventProto_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable - internal_static_org_apache_hadoop_hive_ql_hooks_proto_HiveHookEventProto_fieldAccessorTable; + private static final com.google.protobuf.Descriptors.Descriptor + internal_static_MapFieldEntry_descriptor; + private static final + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable + internal_static_MapFieldEntry_fieldAccessorTable; + private static final com.google.protobuf.Descriptors.Descriptor + internal_static_HiveHookEventProto_descriptor; + private static final + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable + internal_static_HiveHookEventProto_fieldAccessorTable; public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { return descriptor; } - private static com.google.protobuf.Descriptors.FileDescriptor + private static com.google.protobuf.Descriptors.FileDescriptor descriptor; static { java.lang.String[] descriptorData = { - "\n\020HiveEvents.proto\022%org.apache.hadoop.hi" + - "ve.ql.hooks.proto\"+\n\rMapFieldEntry\022\013\n\003ke" + - "y\030\001 \001(\t\022\r\n\005value\030\002 \001(\t\"\241\002\n\022HiveHookEvent" + - "Proto\022\021\n\teventType\030\001 \001(\t\022\023\n\013hiveQueryId\030" + - "\002 \001(\t\022\021\n\ttimestamp\030\003 \001(\003\022\025\n\rexecutionMod" + - "e\030\004 \001(\t\022\023\n\013requestUser\030\005 \001(\t\022\r\n\005queue\030\006 " + - "\001(\t\022\014\n\004user\030\007 \001(\t\022\023\n\013operationId\030\010 \001(\t\022\025" + - "\n\rtablesWritten\030\t \003(\t\022\022\n\ntablesRead\030\n \003(" + - "\t\022G\n\totherInfo\0302 \003(\01324.org.apache.hadoop" + - ".hive.ql.hooks.proto.MapFieldEntryB7\n%or", - "g.apache.hadoop.hive.ql.hooks.protoB\016Hiv" + - "eHookEvents" + "\n\020HiveEvents.proto\"+\n\rMapFieldEntry\022\013\n\003k" + + "ey\030\001 \001(\t\022\r\n\005value\030\002 \001(\t\"\373\001\n\022HiveHookEven" + + "tProto\022\021\n\teventType\030\001 \001(\t\022\023\n\013hiveQueryId" + + "\030\002 \001(\t\022\021\n\ttimestamp\030\003 \001(\003\022\025\n\rexecutionMo" + + "de\030\004 \001(\t\022\023\n\013requestUser\030\005 \001(\t\022\r\n\005queue\030\006" + + " \001(\t\022\014\n\004user\030\007 \001(\t\022\023\n\013operationId\030\010 \001(\t\022" + + "\025\n\rtablesWritten\030\t \003(\t\022\022\n\ntablesRead\030\n \003" + + "(\t\022!\n\totherInfo\0302 \003(\0132\016.MapFieldEntryB7\n" + + "%org.apache.hadoop.hive.ql.hooks.protoB\016" + + "HiveHookEvents" }; - com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = - new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { - public com.google.protobuf.ExtensionRegistry assignDescriptors( - com.google.protobuf.Descriptors.FileDescriptor root) { - descriptor = root; - internal_static_org_apache_hadoop_hive_ql_hooks_proto_MapFieldEntry_descriptor = - getDescriptor().getMessageTypes().get(0); - internal_static_org_apache_hadoop_hive_ql_hooks_proto_MapFieldEntry_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_org_apache_hadoop_hive_ql_hooks_proto_MapFieldEntry_descriptor, - new java.lang.String[] { "Key", "Value", }); - internal_static_org_apache_hadoop_hive_ql_hooks_proto_HiveHookEventProto_descriptor = - getDescriptor().getMessageTypes().get(1); - internal_static_org_apache_hadoop_hive_ql_hooks_proto_HiveHookEventProto_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_org_apache_hadoop_hive_ql_hooks_proto_HiveHookEventProto_descriptor, - new java.lang.String[] { "EventType", "HiveQueryId", "Timestamp", "ExecutionMode", "RequestUser", "Queue", "User", "OperationId", "TablesWritten", "TablesRead", "OtherInfo", }); - return null; - } - }; - com.google.protobuf.Descriptors.FileDescriptor + descriptor = com.google.protobuf.Descriptors.FileDescriptor .internalBuildGeneratedFileFrom(descriptorData, new com.google.protobuf.Descriptors.FileDescriptor[] { - }, assigner); + }); + internal_static_MapFieldEntry_descriptor = + getDescriptor().getMessageTypes().get(0); + internal_static_MapFieldEntry_fieldAccessorTable = new + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_MapFieldEntry_descriptor, + new java.lang.String[] { "Key", "Value", }); + internal_static_HiveHookEventProto_descriptor = + getDescriptor().getMessageTypes().get(1); + internal_static_HiveHookEventProto_fieldAccessorTable = new + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_HiveHookEventProto_descriptor, + new java.lang.String[] { "EventType", "HiveQueryId", "Timestamp", "ExecutionMode", "RequestUser", "Queue", "User", "OperationId", "TablesWritten", "TablesRead", "OtherInfo", }); } // @@protoc_insertion_point(outer_class_scope) diff --git a/ql/src/gen/protobuf/gen-test/org/apache/hadoop/hive/ql/io/protobuf/SampleProtos.java b/ql/src/gen/protobuf/gen-test/org/apache/hadoop/hive/ql/io/protobuf/SampleProtos.java index ac75608439d0..0931fac4eca6 100644 --- a/ql/src/gen/protobuf/gen-test/org/apache/hadoop/hive/ql/io/protobuf/SampleProtos.java +++ b/ql/src/gen/protobuf/gen-test/org/apache/hadoop/hive/ql/io/protobuf/SampleProtos.java @@ -5,38 +5,49 @@ public final class SampleProtos { private SampleProtos() {} + public static void registerAllExtensions( + com.google.protobuf.ExtensionRegistryLite registry) { + } + public static void registerAllExtensions( com.google.protobuf.ExtensionRegistry registry) { + registerAllExtensions( + (com.google.protobuf.ExtensionRegistryLite) registry); } - public interface MapFieldEntryOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface MapFieldEntryOrBuilder extends + // @@protoc_insertion_point(interface_extends:MapFieldEntry) + com.google.protobuf.MessageOrBuilder { - // optional string key = 1; /** * optional string key = 1; + * @return Whether the key field is set. */ boolean hasKey(); /** * optional string key = 1; + * @return The key. */ java.lang.String getKey(); /** * optional string key = 1; + * @return The bytes for key. */ com.google.protobuf.ByteString getKeyBytes(); - // optional string value = 2; /** * optional string value = 2; + * @return Whether the value field is set. */ boolean hasValue(); /** * optional string value = 2; + * @return The value. */ java.lang.String getValue(); /** * optional string value = 2; + * @return The bytes for value. */ com.google.protobuf.ByteString getValueBytes(); @@ -45,35 +56,39 @@ public interface MapFieldEntryOrBuilder * Protobuf type {@code MapFieldEntry} */ public static final class MapFieldEntry extends - com.google.protobuf.GeneratedMessage - implements MapFieldEntryOrBuilder { + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:MapFieldEntry) + MapFieldEntryOrBuilder { + private static final long serialVersionUID = 0L; // Use MapFieldEntry.newBuilder() to construct. - private MapFieldEntry(com.google.protobuf.GeneratedMessage.Builder builder) { + private MapFieldEntry(com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private MapFieldEntry(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final MapFieldEntry defaultInstance; - public static MapFieldEntry getDefaultInstance() { - return defaultInstance; + private MapFieldEntry() { + key_ = ""; + value_ = ""; } - public MapFieldEntry getDefaultInstanceForType() { - return defaultInstance; + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance( + UnusedPrivateParameter unused) { + return new MapFieldEntry(); } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private MapFieldEntry( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); @@ -85,21 +100,23 @@ private MapFieldEntry( case 0: done = true; break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } case 10: { + com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000001; - key_ = input.readBytes(); + key_ = bs; break; } case 18: { + com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000002; - value_ = input.readBytes(); + value_ = bs; + break; + } + default: { + if (!parseUnknownField( + input, unknownFields, extensionRegistry, tag)) { + done = true; + } break; } } @@ -108,7 +125,7 @@ private MapFieldEntry( throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); @@ -119,41 +136,30 @@ private MapFieldEntry( return org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.internal_static_MapFieldEntry_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.internal_static_MapFieldEntry_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry.class, org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public MapFieldEntry parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new MapFieldEntry(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - private int bitField0_; - // optional string key = 1; public static final int KEY_FIELD_NUMBER = 1; - private java.lang.Object key_; + private volatile java.lang.Object key_; /** * optional string key = 1; + * @return Whether the key field is set. */ + @java.lang.Override public boolean hasKey() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional string key = 1; + * @return The key. */ + @java.lang.Override public java.lang.String getKey() { java.lang.Object ref = key_; if (ref instanceof java.lang.String) { @@ -170,7 +176,9 @@ public java.lang.String getKey() { } /** * optional string key = 1; + * @return The bytes for key. */ + @java.lang.Override public com.google.protobuf.ByteString getKeyBytes() { java.lang.Object ref = key_; @@ -185,18 +193,21 @@ public java.lang.String getKey() { } } - // optional string value = 2; public static final int VALUE_FIELD_NUMBER = 2; - private java.lang.Object value_; + private volatile java.lang.Object value_; /** * optional string value = 2; + * @return Whether the value field is set. */ + @java.lang.Override public boolean hasValue() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * optional string value = 2; + * @return The value. */ + @java.lang.Override public java.lang.String getValue() { java.lang.Object ref = value_; if (ref instanceof java.lang.String) { @@ -213,7 +224,9 @@ public java.lang.String getValue() { } /** * optional string value = 2; + * @return The bytes for value. */ + @java.lang.Override public com.google.protobuf.ByteString getValueBytes() { java.lang.Object ref = value_; @@ -228,57 +241,101 @@ public java.lang.String getValue() { } } - private void initFields() { - key_ = ""; - value_ = ""; - } private byte memoizedIsInitialized = -1; + @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } + @java.lang.Override public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); - if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeBytes(1, getKeyBytes()); + if (((bitField0_ & 0x00000001) != 0)) { + com.google.protobuf.GeneratedMessageV3.writeString(output, 1, key_); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - output.writeBytes(2, getValueBytes()); + if (((bitField0_ & 0x00000002) != 0)) { + com.google.protobuf.GeneratedMessageV3.writeString(output, 2, value_); } - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; + @java.lang.Override public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(1, getKeyBytes()); + if (((bitField0_ & 0x00000001) != 0)) { + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, key_); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(2, getValueBytes()); + if (((bitField0_ & 0x00000002) != 0)) { + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(2, value_); } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } - private static final long serialVersionUID = 0L; @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); + public boolean equals(final java.lang.Object obj) { + if (obj == this) { + return true; + } + if (!(obj instanceof org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry)) { + return super.equals(obj); + } + org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry other = (org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry) obj; + + if (hasKey() != other.hasKey()) return false; + if (hasKey()) { + if (!getKey() + .equals(other.getKey())) return false; + } + if (hasValue() != other.hasValue()) return false; + if (hasValue()) { + if (!getValue() + .equals(other.getValue())) return false; + } + if (!unknownFields.equals(other.unknownFields)) return false; + return true; + } + + @java.lang.Override + public int hashCode() { + if (memoizedHashCode != 0) { + return memoizedHashCode; + } + int hash = 41; + hash = (19 * hash) + getDescriptor().hashCode(); + if (hasKey()) { + hash = (37 * hash) + KEY_FIELD_NUMBER; + hash = (53 * hash) + getKey().hashCode(); + } + if (hasValue()) { + hash = (37 * hash) + VALUE_FIELD_NUMBER; + hash = (53 * hash) + getValue().hashCode(); + } + hash = (29 * hash) + unknownFields.hashCode(); + memoizedHashCode = hash; + return hash; } + public static org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry parseFrom( + java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry parseFrom( + java.nio.ByteBuffer data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } public static org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { @@ -302,46 +359,59 @@ public static org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry p } public static org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } + @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -349,14 +419,16 @@ protected Builder newBuilderForType( * Protobuf type {@code MapFieldEntry} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntryOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder implements + // @@protoc_insertion_point(builder_implements:MapFieldEntry) + org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntryOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.internal_static_MapFieldEntry_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.internal_static_MapFieldEntry_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -369,18 +441,16 @@ private Builder() { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { } } - private static Builder create() { - return new Builder(); - } - + @java.lang.Override public Builder clear() { super.clear(); key_ = ""; @@ -390,19 +460,18 @@ public Builder clear() { return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - + @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.internal_static_MapFieldEntry_descriptor; } + @java.lang.Override public org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry getDefaultInstanceForType() { return org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry.getDefaultInstance(); } + @java.lang.Override public org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry build() { org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry result = buildPartial(); if (!result.isInitialized()) { @@ -411,15 +480,16 @@ public org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry build() return result; } + @java.lang.Override public org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry buildPartial() { org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry result = new org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { + if (((from_bitField0_ & 0x00000001) != 0)) { to_bitField0_ |= 0x00000001; } result.key_ = key_; - if (((from_bitField0_ & 0x00000002) == 0x00000002)) { + if (((from_bitField0_ & 0x00000002) != 0)) { to_bitField0_ |= 0x00000002; } result.value_ = value_; @@ -428,6 +498,39 @@ public org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry buildPar return result; } + @java.lang.Override + public Builder clone() { + return super.clone(); + } + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.setField(field, value); + } + @java.lang.Override + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + @java.lang.Override + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.addRepeatedField(field, value); + } + @java.lang.Override public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry) { return mergeFrom((org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry)other); @@ -449,14 +552,17 @@ public Builder mergeFrom(org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapF value_ = other.value_; onChanged(); } - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } + @java.lang.Override public final boolean isInitialized() { return true; } + @java.lang.Override public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) @@ -466,7 +572,7 @@ public Builder mergeFrom( parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -476,23 +582,27 @@ public Builder mergeFrom( } private int bitField0_; - // optional string key = 1; private java.lang.Object key_ = ""; /** * optional string key = 1; + * @return Whether the key field is set. */ public boolean hasKey() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional string key = 1; + * @return The key. */ public java.lang.String getKey() { java.lang.Object ref = key_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - key_ = s; + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + key_ = s; + } return s; } else { return (java.lang.String) ref; @@ -500,6 +610,7 @@ public java.lang.String getKey() { } /** * optional string key = 1; + * @return The bytes for key. */ public com.google.protobuf.ByteString getKeyBytes() { @@ -516,6 +627,8 @@ public java.lang.String getKey() { } /** * optional string key = 1; + * @param value The key to set. + * @return This builder for chaining. */ public Builder setKey( java.lang.String value) { @@ -529,6 +642,7 @@ public Builder setKey( } /** * optional string key = 1; + * @return This builder for chaining. */ public Builder clearKey() { bitField0_ = (bitField0_ & ~0x00000001); @@ -538,6 +652,8 @@ public Builder clearKey() { } /** * optional string key = 1; + * @param value The bytes for key to set. + * @return This builder for chaining. */ public Builder setKeyBytes( com.google.protobuf.ByteString value) { @@ -550,23 +666,27 @@ public Builder setKeyBytes( return this; } - // optional string value = 2; private java.lang.Object value_ = ""; /** * optional string value = 2; + * @return Whether the value field is set. */ public boolean hasValue() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * optional string value = 2; + * @return The value. */ public java.lang.String getValue() { java.lang.Object ref = value_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - value_ = s; + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + value_ = s; + } return s; } else { return (java.lang.String) ref; @@ -574,6 +694,7 @@ public java.lang.String getValue() { } /** * optional string value = 2; + * @return The bytes for value. */ public com.google.protobuf.ByteString getValueBytes() { @@ -590,6 +711,8 @@ public java.lang.String getValue() { } /** * optional string value = 2; + * @param value The value to set. + * @return This builder for chaining. */ public Builder setValue( java.lang.String value) { @@ -603,6 +726,7 @@ public Builder setValue( } /** * optional string value = 2; + * @return This builder for chaining. */ public Builder clearValue() { bitField0_ = (bitField0_ & ~0x00000002); @@ -612,6 +736,8 @@ public Builder clearValue() { } /** * optional string value = 2; + * @param value The bytes for value to set. + * @return This builder for chaining. */ public Builder setValueBytes( com.google.protobuf.ByteString value) { @@ -623,22 +749,63 @@ public Builder setValueBytes( onChanged(); return this; } + @java.lang.Override + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:MapFieldEntry) } + // @@protoc_insertion_point(class_scope:MapFieldEntry) + private static final org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry DEFAULT_INSTANCE; static { - defaultInstance = new MapFieldEntry(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry(); + } + + public static org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser + PARSER = new com.google.protobuf.AbstractParser() { + @java.lang.Override + public MapFieldEntry parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new MapFieldEntry(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry getDefaultInstanceForType() { + return DEFAULT_INSTANCE; } - // @@protoc_insertion_point(class_scope:MapFieldEntry) } - public interface Mesg1OrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface Mesg1OrBuilder extends + // @@protoc_insertion_point(interface_extends:Mesg1) + com.google.protobuf.MessageOrBuilder { - // repeated .MapFieldEntry anotherMap = 1; /** * repeated .MapFieldEntry anotherMap = 1; */ @@ -663,13 +830,14 @@ public interface Mesg1OrBuilder org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntryOrBuilder getAnotherMapOrBuilder( int index); - // optional .MapFieldEntry noMap = 2; /** * optional .MapFieldEntry noMap = 2; + * @return Whether the noMap field is set. */ boolean hasNoMap(); /** * optional .MapFieldEntry noMap = 2; + * @return The noMap. */ org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry getNoMap(); /** @@ -677,17 +845,20 @@ org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntryOrBuilder getAno */ org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntryOrBuilder getNoMapOrBuilder(); - // repeated int32 intList = 3; /** * repeated int32 intList = 3; + * @return A list containing the intList. */ java.util.List getIntListList(); /** * repeated int32 intList = 3; + * @return The count of intList. */ int getIntListCount(); /** * repeated int32 intList = 3; + * @param index The index of the element to return. + * @return The intList at the given index. */ int getIntList(int index); } @@ -695,35 +866,39 @@ org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntryOrBuilder getAno * Protobuf type {@code Mesg1} */ public static final class Mesg1 extends - com.google.protobuf.GeneratedMessage - implements Mesg1OrBuilder { + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:Mesg1) + Mesg1OrBuilder { + private static final long serialVersionUID = 0L; // Use Mesg1.newBuilder() to construct. - private Mesg1(com.google.protobuf.GeneratedMessage.Builder builder) { + private Mesg1(com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private Mesg1(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final Mesg1 defaultInstance; - public static Mesg1 getDefaultInstance() { - return defaultInstance; + private Mesg1() { + anotherMap_ = java.util.Collections.emptyList(); + intList_ = emptyIntList(); } - public Mesg1 getDefaultInstanceForType() { - return defaultInstance; + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance( + UnusedPrivateParameter unused) { + return new Mesg1(); } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private Mesg1( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); @@ -735,24 +910,18 @@ private Mesg1( case 0: done = true; break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } case 10: { - if (!((mutable_bitField0_ & 0x00000001) == 0x00000001)) { + if (!((mutable_bitField0_ & 0x00000001) != 0)) { anotherMap_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000001; } - anotherMap_.add(input.readMessage(org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry.PARSER, extensionRegistry)); + anotherMap_.add( + input.readMessage(org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry.PARSER, extensionRegistry)); break; } case 18: { org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry.Builder subBuilder = null; - if (((bitField0_ & 0x00000001) == 0x00000001)) { + if (((bitField0_ & 0x00000001) != 0)) { subBuilder = noMap_.toBuilder(); } noMap_ = input.readMessage(org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry.PARSER, extensionRegistry); @@ -764,39 +933,46 @@ private Mesg1( break; } case 24: { - if (!((mutable_bitField0_ & 0x00000004) == 0x00000004)) { - intList_ = new java.util.ArrayList(); + if (!((mutable_bitField0_ & 0x00000004) != 0)) { + intList_ = newIntList(); mutable_bitField0_ |= 0x00000004; } - intList_.add(input.readInt32()); + intList_.addInt(input.readInt32()); break; } case 26: { int length = input.readRawVarint32(); int limit = input.pushLimit(length); - if (!((mutable_bitField0_ & 0x00000004) == 0x00000004) && input.getBytesUntilLimit() > 0) { - intList_ = new java.util.ArrayList(); + if (!((mutable_bitField0_ & 0x00000004) != 0) && input.getBytesUntilLimit() > 0) { + intList_ = newIntList(); mutable_bitField0_ |= 0x00000004; } while (input.getBytesUntilLimit() > 0) { - intList_.add(input.readInt32()); + intList_.addInt(input.readInt32()); } input.popLimit(limit); break; } + default: { + if (!parseUnknownField( + input, unknownFields, extensionRegistry, tag)) { + done = true; + } + break; + } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { - if (((mutable_bitField0_ & 0x00000001) == 0x00000001)) { + if (((mutable_bitField0_ & 0x00000001) != 0)) { anotherMap_ = java.util.Collections.unmodifiableList(anotherMap_); } - if (((mutable_bitField0_ & 0x00000004) == 0x00000004)) { - intList_ = java.util.Collections.unmodifiableList(intList_); + if (((mutable_bitField0_ & 0x00000004) != 0)) { + intList_.makeImmutable(); // C } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); @@ -807,41 +983,28 @@ private Mesg1( return org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.internal_static_Mesg1_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.internal_static_Mesg1_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1.class, org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public Mesg1 parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new Mesg1(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - private int bitField0_; - // repeated .MapFieldEntry anotherMap = 1; public static final int ANOTHERMAP_FIELD_NUMBER = 1; private java.util.List anotherMap_; /** * repeated .MapFieldEntry anotherMap = 1; */ + @java.lang.Override public java.util.List getAnotherMapList() { return anotherMap_; } /** * repeated .MapFieldEntry anotherMap = 1; */ + @java.lang.Override public java.util.List getAnotherMapOrBuilderList() { return anotherMap_; @@ -849,100 +1012,108 @@ public java.util.Listrepeated .MapFieldEntry anotherMap = 1; */ + @java.lang.Override public int getAnotherMapCount() { return anotherMap_.size(); } /** * repeated .MapFieldEntry anotherMap = 1; */ + @java.lang.Override public org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry getAnotherMap(int index) { return anotherMap_.get(index); } /** * repeated .MapFieldEntry anotherMap = 1; */ + @java.lang.Override public org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntryOrBuilder getAnotherMapOrBuilder( int index) { return anotherMap_.get(index); } - // optional .MapFieldEntry noMap = 2; public static final int NOMAP_FIELD_NUMBER = 2; private org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry noMap_; /** * optional .MapFieldEntry noMap = 2; + * @return Whether the noMap field is set. */ + @java.lang.Override public boolean hasNoMap() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional .MapFieldEntry noMap = 2; + * @return The noMap. */ + @java.lang.Override public org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry getNoMap() { - return noMap_; + return noMap_ == null ? org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry.getDefaultInstance() : noMap_; } /** * optional .MapFieldEntry noMap = 2; */ + @java.lang.Override public org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntryOrBuilder getNoMapOrBuilder() { - return noMap_; + return noMap_ == null ? org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry.getDefaultInstance() : noMap_; } - // repeated int32 intList = 3; public static final int INTLIST_FIELD_NUMBER = 3; - private java.util.List intList_; + private com.google.protobuf.Internal.IntList intList_; /** * repeated int32 intList = 3; + * @return A list containing the intList. */ + @java.lang.Override public java.util.List getIntListList() { return intList_; } /** * repeated int32 intList = 3; + * @return The count of intList. */ public int getIntListCount() { return intList_.size(); } /** * repeated int32 intList = 3; + * @param index The index of the element to return. + * @return The intList at the given index. */ public int getIntList(int index) { - return intList_.get(index); + return intList_.getInt(index); } - private void initFields() { - anotherMap_ = java.util.Collections.emptyList(); - noMap_ = org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry.getDefaultInstance(); - intList_ = java.util.Collections.emptyList(); - } private byte memoizedIsInitialized = -1; + @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } + @java.lang.Override public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); for (int i = 0; i < anotherMap_.size(); i++) { output.writeMessage(1, anotherMap_.get(i)); } - if (((bitField0_ & 0x00000001) == 0x00000001)) { - output.writeMessage(2, noMap_); + if (((bitField0_ & 0x00000001) != 0)) { + output.writeMessage(2, getNoMap()); } for (int i = 0; i < intList_.size(); i++) { - output.writeInt32(3, intList_.get(i)); + output.writeInt32(3, intList_.getInt(i)); } - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; + @java.lang.Override public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; @@ -950,31 +1121,82 @@ public int getSerializedSize() { size += com.google.protobuf.CodedOutputStream .computeMessageSize(1, anotherMap_.get(i)); } - if (((bitField0_ & 0x00000001) == 0x00000001)) { + if (((bitField0_ & 0x00000001) != 0)) { size += com.google.protobuf.CodedOutputStream - .computeMessageSize(2, noMap_); + .computeMessageSize(2, getNoMap()); } { int dataSize = 0; for (int i = 0; i < intList_.size(); i++) { dataSize += com.google.protobuf.CodedOutputStream - .computeInt32SizeNoTag(intList_.get(i)); + .computeInt32SizeNoTag(intList_.getInt(i)); } size += dataSize; size += 1 * getIntListList().size(); } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } - private static final long serialVersionUID = 0L; @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); + public boolean equals(final java.lang.Object obj) { + if (obj == this) { + return true; + } + if (!(obj instanceof org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1)) { + return super.equals(obj); + } + org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1 other = (org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1) obj; + + if (!getAnotherMapList() + .equals(other.getAnotherMapList())) return false; + if (hasNoMap() != other.hasNoMap()) return false; + if (hasNoMap()) { + if (!getNoMap() + .equals(other.getNoMap())) return false; + } + if (!getIntListList() + .equals(other.getIntListList())) return false; + if (!unknownFields.equals(other.unknownFields)) return false; + return true; + } + + @java.lang.Override + public int hashCode() { + if (memoizedHashCode != 0) { + return memoizedHashCode; + } + int hash = 41; + hash = (19 * hash) + getDescriptor().hashCode(); + if (getAnotherMapCount() > 0) { + hash = (37 * hash) + ANOTHERMAP_FIELD_NUMBER; + hash = (53 * hash) + getAnotherMapList().hashCode(); + } + if (hasNoMap()) { + hash = (37 * hash) + NOMAP_FIELD_NUMBER; + hash = (53 * hash) + getNoMap().hashCode(); + } + if (getIntListCount() > 0) { + hash = (37 * hash) + INTLIST_FIELD_NUMBER; + hash = (53 * hash) + getIntListList().hashCode(); + } + hash = (29 * hash) + unknownFields.hashCode(); + memoizedHashCode = hash; + return hash; } + public static org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1 parseFrom( + java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1 parseFrom( + java.nio.ByteBuffer data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } public static org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1 parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { @@ -998,46 +1220,59 @@ public static org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1 parseFrom } public static org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1 parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1 parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1 parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1 parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1 parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1 parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } + @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1 prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -1045,14 +1280,16 @@ protected Builder newBuilderForType( * Protobuf type {@code Mesg1} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1OrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder implements + // @@protoc_insertion_point(builder_implements:Mesg1) + org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1OrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.internal_static_Mesg1_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.internal_static_Mesg1_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -1065,20 +1302,18 @@ private Builder() { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { getAnotherMapFieldBuilder(); getNoMapFieldBuilder(); } } - private static Builder create() { - return new Builder(); - } - + @java.lang.Override public Builder clear() { super.clear(); if (anotherMapBuilder_ == null) { @@ -1088,29 +1323,28 @@ public Builder clear() { anotherMapBuilder_.clear(); } if (noMapBuilder_ == null) { - noMap_ = org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry.getDefaultInstance(); + noMap_ = null; } else { noMapBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000002); - intList_ = java.util.Collections.emptyList(); + intList_ = emptyIntList(); bitField0_ = (bitField0_ & ~0x00000004); return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - + @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.internal_static_Mesg1_descriptor; } + @java.lang.Override public org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1 getDefaultInstanceForType() { return org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1.getDefaultInstance(); } + @java.lang.Override public org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1 build() { org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1 result = buildPartial(); if (!result.isInitialized()) { @@ -1119,12 +1353,13 @@ public org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1 build() { return result; } + @java.lang.Override public org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1 buildPartial() { org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1 result = new org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (anotherMapBuilder_ == null) { - if (((bitField0_ & 0x00000001) == 0x00000001)) { + if (((bitField0_ & 0x00000001) != 0)) { anotherMap_ = java.util.Collections.unmodifiableList(anotherMap_); bitField0_ = (bitField0_ & ~0x00000001); } @@ -1132,16 +1367,16 @@ public org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1 buildPartial() { } else { result.anotherMap_ = anotherMapBuilder_.build(); } - if (((from_bitField0_ & 0x00000002) == 0x00000002)) { + if (((from_bitField0_ & 0x00000002) != 0)) { + if (noMapBuilder_ == null) { + result.noMap_ = noMap_; + } else { + result.noMap_ = noMapBuilder_.build(); + } to_bitField0_ |= 0x00000001; } - if (noMapBuilder_ == null) { - result.noMap_ = noMap_; - } else { - result.noMap_ = noMapBuilder_.build(); - } - if (((bitField0_ & 0x00000004) == 0x00000004)) { - intList_ = java.util.Collections.unmodifiableList(intList_); + if (((bitField0_ & 0x00000004) != 0)) { + intList_.makeImmutable(); bitField0_ = (bitField0_ & ~0x00000004); } result.intList_ = intList_; @@ -1150,6 +1385,39 @@ public org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1 buildPartial() { return result; } + @java.lang.Override + public Builder clone() { + return super.clone(); + } + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.setField(field, value); + } + @java.lang.Override + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + @java.lang.Override + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.addRepeatedField(field, value); + } + @java.lang.Override public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1) { return mergeFrom((org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1)other); @@ -1180,7 +1448,7 @@ public Builder mergeFrom(org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg anotherMap_ = other.anotherMap_; bitField0_ = (bitField0_ & ~0x00000001); anotherMapBuilder_ = - com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? + com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getAnotherMapFieldBuilder() : null; } else { anotherMapBuilder_.addAllMessages(other.anotherMap_); @@ -1200,14 +1468,17 @@ public Builder mergeFrom(org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg } onChanged(); } - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } + @java.lang.Override public final boolean isInitialized() { return true; } + @java.lang.Override public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) @@ -1217,7 +1488,7 @@ public Builder mergeFrom( parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -1227,17 +1498,16 @@ public Builder mergeFrom( } private int bitField0_; - // repeated .MapFieldEntry anotherMap = 1; private java.util.List anotherMap_ = java.util.Collections.emptyList(); private void ensureAnotherMapIsMutable() { - if (!((bitField0_ & 0x00000001) == 0x00000001)) { + if (!((bitField0_ & 0x00000001) != 0)) { anotherMap_ = new java.util.ArrayList(anotherMap_); bitField0_ |= 0x00000001; } } - private com.google.protobuf.RepeatedFieldBuilder< + private com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry, org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry.Builder, org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntryOrBuilder> anotherMapBuilder_; /** @@ -1369,7 +1639,8 @@ public Builder addAllAnotherMap( java.lang.Iterable values) { if (anotherMapBuilder_ == null) { ensureAnotherMapIsMutable(); - super.addAll(values, anotherMap_); + com.google.protobuf.AbstractMessageLite.Builder.addAll( + values, anotherMap_); onChanged(); } else { anotherMapBuilder_.addAllMessages(values); @@ -1452,14 +1723,14 @@ public org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry.Builder getAnotherMapBuilderList() { return getAnotherMapFieldBuilder().getBuilderList(); } - private com.google.protobuf.RepeatedFieldBuilder< + private com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry, org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry.Builder, org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntryOrBuilder> getAnotherMapFieldBuilder() { if (anotherMapBuilder_ == null) { - anotherMapBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< + anotherMapBuilder_ = new com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry, org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry.Builder, org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntryOrBuilder>( anotherMap_, - ((bitField0_ & 0x00000001) == 0x00000001), + ((bitField0_ & 0x00000001) != 0), getParentForChildren(), isClean()); anotherMap_ = null; @@ -1467,22 +1738,23 @@ public org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry.Builder return anotherMapBuilder_; } - // optional .MapFieldEntry noMap = 2; - private org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry noMap_ = org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry.getDefaultInstance(); - private com.google.protobuf.SingleFieldBuilder< + private org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry noMap_; + private com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry, org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry.Builder, org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntryOrBuilder> noMapBuilder_; /** * optional .MapFieldEntry noMap = 2; + * @return Whether the noMap field is set. */ public boolean hasNoMap() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * optional .MapFieldEntry noMap = 2; + * @return The noMap. */ public org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry getNoMap() { if (noMapBuilder_ == null) { - return noMap_; + return noMap_ == null ? org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry.getDefaultInstance() : noMap_; } else { return noMapBuilder_.getMessage(); } @@ -1522,7 +1794,8 @@ public Builder setNoMap( */ public Builder mergeNoMap(org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry value) { if (noMapBuilder_ == null) { - if (((bitField0_ & 0x00000002) == 0x00000002) && + if (((bitField0_ & 0x00000002) != 0) && + noMap_ != null && noMap_ != org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry.getDefaultInstance()) { noMap_ = org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry.newBuilder(noMap_).mergeFrom(value).buildPartial(); @@ -1541,7 +1814,7 @@ public Builder mergeNoMap(org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Map */ public Builder clearNoMap() { if (noMapBuilder_ == null) { - noMap_ = org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry.getDefaultInstance(); + noMap_ = null; onChanged(); } else { noMapBuilder_.clear(); @@ -1564,19 +1837,20 @@ public org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntryOrBuilder if (noMapBuilder_ != null) { return noMapBuilder_.getMessageOrBuilder(); } else { - return noMap_; + return noMap_ == null ? + org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry.getDefaultInstance() : noMap_; } } /** * optional .MapFieldEntry noMap = 2; */ - private com.google.protobuf.SingleFieldBuilder< + private com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry, org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry.Builder, org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntryOrBuilder> getNoMapFieldBuilder() { if (noMapBuilder_ == null) { - noMapBuilder_ = new com.google.protobuf.SingleFieldBuilder< + noMapBuilder_ = new com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry, org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry.Builder, org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntryOrBuilder>( - noMap_, + getNoMap(), getParentForChildren(), isClean()); noMap_ = null; @@ -1584,242 +1858,312 @@ public org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntryOrBuilder return noMapBuilder_; } - // repeated int32 intList = 3; - private java.util.List intList_ = java.util.Collections.emptyList(); + private com.google.protobuf.Internal.IntList intList_ = emptyIntList(); private void ensureIntListIsMutable() { - if (!((bitField0_ & 0x00000004) == 0x00000004)) { - intList_ = new java.util.ArrayList(intList_); + if (!((bitField0_ & 0x00000004) != 0)) { + intList_ = mutableCopy(intList_); bitField0_ |= 0x00000004; } } /** * repeated int32 intList = 3; + * @return A list containing the intList. */ public java.util.List getIntListList() { - return java.util.Collections.unmodifiableList(intList_); + return ((bitField0_ & 0x00000004) != 0) ? + java.util.Collections.unmodifiableList(intList_) : intList_; } /** * repeated int32 intList = 3; + * @return The count of intList. */ public int getIntListCount() { return intList_.size(); } /** * repeated int32 intList = 3; + * @param index The index of the element to return. + * @return The intList at the given index. */ public int getIntList(int index) { - return intList_.get(index); + return intList_.getInt(index); } /** * repeated int32 intList = 3; + * @param index The index to set the value at. + * @param value The intList to set. + * @return This builder for chaining. */ public Builder setIntList( int index, int value) { ensureIntListIsMutable(); - intList_.set(index, value); + intList_.setInt(index, value); onChanged(); return this; } /** * repeated int32 intList = 3; + * @param value The intList to add. + * @return This builder for chaining. */ public Builder addIntList(int value) { ensureIntListIsMutable(); - intList_.add(value); + intList_.addInt(value); onChanged(); return this; } /** * repeated int32 intList = 3; + * @param values The intList to add. + * @return This builder for chaining. */ public Builder addAllIntList( java.lang.Iterable values) { ensureIntListIsMutable(); - super.addAll(values, intList_); + com.google.protobuf.AbstractMessageLite.Builder.addAll( + values, intList_); onChanged(); return this; } /** * repeated int32 intList = 3; + * @return This builder for chaining. */ public Builder clearIntList() { - intList_ = java.util.Collections.emptyList(); + intList_ = emptyIntList(); bitField0_ = (bitField0_ & ~0x00000004); onChanged(); return this; } + @java.lang.Override + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:Mesg1) } + // @@protoc_insertion_point(class_scope:Mesg1) + private static final org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1 DEFAULT_INSTANCE; static { - defaultInstance = new Mesg1(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1(); + } + + public static org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1 getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser + PARSER = new com.google.protobuf.AbstractParser() { + @java.lang.Override + public Mesg1 parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new Mesg1(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1 getDefaultInstanceForType() { + return DEFAULT_INSTANCE; } - // @@protoc_insertion_point(class_scope:Mesg1) } - public interface AllTypesOrBuilder - extends com.google.protobuf.MessageOrBuilder { + public interface AllTypesOrBuilder extends + // @@protoc_insertion_point(interface_extends:AllTypes) + com.google.protobuf.MessageOrBuilder { - // optional double doubleType = 1; /** * optional double doubleType = 1; + * @return Whether the doubleType field is set. */ boolean hasDoubleType(); /** * optional double doubleType = 1; + * @return The doubleType. */ double getDoubleType(); - // optional float floatType = 2; /** * optional float floatType = 2; + * @return Whether the floatType field is set. */ boolean hasFloatType(); /** * optional float floatType = 2; + * @return The floatType. */ float getFloatType(); - // optional int32 int32Type = 3; /** * optional int32 int32Type = 3; + * @return Whether the int32Type field is set. */ boolean hasInt32Type(); /** * optional int32 int32Type = 3; + * @return The int32Type. */ int getInt32Type(); - // optional int64 int64Type = 4; /** * optional int64 int64Type = 4; + * @return Whether the int64Type field is set. */ boolean hasInt64Type(); /** * optional int64 int64Type = 4; + * @return The int64Type. */ long getInt64Type(); - // optional uint32 uint32Type = 5; /** * optional uint32 uint32Type = 5; + * @return Whether the uint32Type field is set. */ boolean hasUint32Type(); /** * optional uint32 uint32Type = 5; + * @return The uint32Type. */ int getUint32Type(); - // optional uint64 uint64Type = 6; /** * optional uint64 uint64Type = 6; + * @return Whether the uint64Type field is set. */ boolean hasUint64Type(); /** * optional uint64 uint64Type = 6; + * @return The uint64Type. */ long getUint64Type(); - // optional sint32 sint32Type = 7; /** * optional sint32 sint32Type = 7; + * @return Whether the sint32Type field is set. */ boolean hasSint32Type(); /** * optional sint32 sint32Type = 7; + * @return The sint32Type. */ int getSint32Type(); - // optional sint64 sint64Type = 8; /** * optional sint64 sint64Type = 8; + * @return Whether the sint64Type field is set. */ boolean hasSint64Type(); /** * optional sint64 sint64Type = 8; + * @return The sint64Type. */ long getSint64Type(); - // optional fixed32 fixed32Type = 9; /** * optional fixed32 fixed32Type = 9; + * @return Whether the fixed32Type field is set. */ boolean hasFixed32Type(); /** * optional fixed32 fixed32Type = 9; + * @return The fixed32Type. */ int getFixed32Type(); - // optional fixed64 fixed64Type = 10; /** * optional fixed64 fixed64Type = 10; + * @return Whether the fixed64Type field is set. */ boolean hasFixed64Type(); /** * optional fixed64 fixed64Type = 10; + * @return The fixed64Type. */ long getFixed64Type(); - // optional sfixed32 sfixed32Type = 11; /** * optional sfixed32 sfixed32Type = 11; + * @return Whether the sfixed32Type field is set. */ boolean hasSfixed32Type(); /** * optional sfixed32 sfixed32Type = 11; + * @return The sfixed32Type. */ int getSfixed32Type(); - // optional sfixed64 sfixed64Type = 12; /** * optional sfixed64 sfixed64Type = 12; + * @return Whether the sfixed64Type field is set. */ boolean hasSfixed64Type(); /** * optional sfixed64 sfixed64Type = 12; + * @return The sfixed64Type. */ long getSfixed64Type(); - // optional bool boolType = 13; /** * optional bool boolType = 13; + * @return Whether the boolType field is set. */ boolean hasBoolType(); /** * optional bool boolType = 13; + * @return The boolType. */ boolean getBoolType(); - // optional string stringType = 14; /** * optional string stringType = 14; + * @return Whether the stringType field is set. */ boolean hasStringType(); /** * optional string stringType = 14; + * @return The stringType. */ java.lang.String getStringType(); /** * optional string stringType = 14; + * @return The bytes for stringType. */ com.google.protobuf.ByteString getStringTypeBytes(); - // optional bytes bytesType = 15; /** * optional bytes bytesType = 15; + * @return Whether the bytesType field is set. */ boolean hasBytesType(); /** * optional bytes bytesType = 15; + * @return The bytesType. */ com.google.protobuf.ByteString getBytesType(); - // repeated .MapFieldEntry mapType = 16; /** * repeated .MapFieldEntry mapType = 16; */ @@ -1844,33 +2188,39 @@ public interface AllTypesOrBuilder org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntryOrBuilder getMapTypeOrBuilder( int index); - // repeated string stringListType = 17; /** * repeated string stringListType = 17; + * @return A list containing the stringListType. */ java.util.List - getStringListTypeList(); + getStringListTypeList(); /** * repeated string stringListType = 17; + * @return The count of stringListType. */ int getStringListTypeCount(); /** * repeated string stringListType = 17; + * @param index The index of the element to return. + * @return The stringListType at the given index. */ java.lang.String getStringListType(int index); /** * repeated string stringListType = 17; + * @param index The index of the value to return. + * @return The bytes of the stringListType at the given index. */ com.google.protobuf.ByteString getStringListTypeBytes(int index); - // optional .Mesg1 messageType = 18; /** * optional .Mesg1 messageType = 18; + * @return Whether the messageType field is set. */ boolean hasMessageType(); /** * optional .Mesg1 messageType = 18; + * @return The messageType. */ org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1 getMessageType(); /** @@ -1878,7 +2228,6 @@ org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntryOrBuilder getMap */ org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1OrBuilder getMessageTypeOrBuilder(); - // repeated .Mesg1 messageListType = 19; /** * repeated .Mesg1 messageListType = 19; */ @@ -1903,13 +2252,14 @@ org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntryOrBuilder getMap org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1OrBuilder getMessageListTypeOrBuilder( int index); - // optional .AllTypes.Enum1 enumType = 20; /** * optional .AllTypes.Enum1 enumType = 20; + * @return Whether the enumType field is set. */ boolean hasEnumType(); /** * optional .AllTypes.Enum1 enumType = 20; + * @return The enumType. */ org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes.Enum1 getEnumType(); } @@ -1917,35 +2267,43 @@ org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1OrBuilder getMessageList * Protobuf type {@code AllTypes} */ public static final class AllTypes extends - com.google.protobuf.GeneratedMessage - implements AllTypesOrBuilder { + com.google.protobuf.GeneratedMessageV3 implements + // @@protoc_insertion_point(message_implements:AllTypes) + AllTypesOrBuilder { + private static final long serialVersionUID = 0L; // Use AllTypes.newBuilder() to construct. - private AllTypes(com.google.protobuf.GeneratedMessage.Builder builder) { + private AllTypes(com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); - this.unknownFields = builder.getUnknownFields(); } - private AllTypes(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } - - private static final AllTypes defaultInstance; - public static AllTypes getDefaultInstance() { - return defaultInstance; + private AllTypes() { + stringType_ = ""; + bytesType_ = com.google.protobuf.ByteString.EMPTY; + mapType_ = java.util.Collections.emptyList(); + stringListType_ = com.google.protobuf.LazyStringArrayList.EMPTY; + messageListType_ = java.util.Collections.emptyList(); + enumType_ = 1; } - public AllTypes getDefaultInstanceForType() { - return defaultInstance; + @java.lang.Override + @SuppressWarnings({"unused"}) + protected java.lang.Object newInstance( + UnusedPrivateParameter unused) { + return new AllTypes(); } - private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet - getUnknownFields() { + getUnknownFields() { return this.unknownFields; } private AllTypes( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { - initFields(); + this(); + if (extensionRegistry == null) { + throw new java.lang.NullPointerException(); + } int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); @@ -1957,13 +2315,6 @@ private AllTypes( case 0: done = true; break; - default: { - if (!parseUnknownField(input, unknownFields, - extensionRegistry, tag)) { - done = true; - } - break; - } case 9: { bitField0_ |= 0x00000001; doubleType_ = input.readDouble(); @@ -2030,8 +2381,9 @@ private AllTypes( break; } case 114: { + com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00002000; - stringType_ = input.readBytes(); + stringType_ = bs; break; } case 122: { @@ -2040,24 +2392,26 @@ private AllTypes( break; } case 130: { - if (!((mutable_bitField0_ & 0x00008000) == 0x00008000)) { + if (!((mutable_bitField0_ & 0x00008000) != 0)) { mapType_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00008000; } - mapType_.add(input.readMessage(org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry.PARSER, extensionRegistry)); + mapType_.add( + input.readMessage(org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry.PARSER, extensionRegistry)); break; } case 138: { - if (!((mutable_bitField0_ & 0x00010000) == 0x00010000)) { + com.google.protobuf.ByteString bs = input.readBytes(); + if (!((mutable_bitField0_ & 0x00010000) != 0)) { stringListType_ = new com.google.protobuf.LazyStringArrayList(); mutable_bitField0_ |= 0x00010000; } - stringListType_.add(input.readBytes()); + stringListType_.add(bs); break; } case 146: { org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1.Builder subBuilder = null; - if (((bitField0_ & 0x00008000) == 0x00008000)) { + if (((bitField0_ & 0x00008000) != 0)) { subBuilder = messageType_.toBuilder(); } messageType_ = input.readMessage(org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1.PARSER, extensionRegistry); @@ -2069,21 +2423,30 @@ private AllTypes( break; } case 154: { - if (!((mutable_bitField0_ & 0x00040000) == 0x00040000)) { + if (!((mutable_bitField0_ & 0x00040000) != 0)) { messageListType_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00040000; } - messageListType_.add(input.readMessage(org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1.PARSER, extensionRegistry)); + messageListType_.add( + input.readMessage(org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1.PARSER, extensionRegistry)); break; } case 160: { int rawValue = input.readEnum(); + @SuppressWarnings("deprecation") org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes.Enum1 value = org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes.Enum1.valueOf(rawValue); if (value == null) { unknownFields.mergeVarintField(20, rawValue); } else { bitField0_ |= 0x00010000; - enumType_ = value; + enumType_ = rawValue; + } + break; + } + default: { + if (!parseUnknownField( + input, unknownFields, extensionRegistry, tag)) { + done = true; } break; } @@ -2093,15 +2456,15 @@ private AllTypes( throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( - e.getMessage()).setUnfinishedMessage(this); + e).setUnfinishedMessage(this); } finally { - if (((mutable_bitField0_ & 0x00008000) == 0x00008000)) { + if (((mutable_bitField0_ & 0x00008000) != 0)) { mapType_ = java.util.Collections.unmodifiableList(mapType_); } - if (((mutable_bitField0_ & 0x00010000) == 0x00010000)) { - stringListType_ = new com.google.protobuf.UnmodifiableLazyStringList(stringListType_); + if (((mutable_bitField0_ & 0x00010000) != 0)) { + stringListType_ = stringListType_.getUnmodifiableView(); } - if (((mutable_bitField0_ & 0x00040000) == 0x00040000)) { + if (((mutable_bitField0_ & 0x00040000) != 0)) { messageListType_ = java.util.Collections.unmodifiableList(messageListType_); } this.unknownFields = unknownFields.build(); @@ -2113,28 +2476,14 @@ private AllTypes( return org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.internal_static_AllTypes_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.internal_static_AllTypes_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes.class, org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes.Builder.class); } - public static com.google.protobuf.Parser PARSER = - new com.google.protobuf.AbstractParser() { - public AllTypes parsePartialFrom( - com.google.protobuf.CodedInputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return new AllTypes(input, extensionRegistry); - } - }; - - @java.lang.Override - public com.google.protobuf.Parser getParserForType() { - return PARSER; - } - /** * Protobuf enum {@code AllTypes.Enum1} */ @@ -2143,11 +2492,11 @@ public enum Enum1 /** * VAL1 = 1; */ - VAL1(0, 1), + VAL1(1), /** * VAL2 = 2; */ - VAL2(1, 2), + VAL2(2), ; /** @@ -2160,9 +2509,25 @@ public enum Enum1 public static final int VAL2_VALUE = 2; - public final int getNumber() { return value; } + public final int getNumber() { + return value; + } + /** + * @param value The numeric wire value of the corresponding enum entry. + * @return The enum associated with the given numeric wire value. + * @deprecated Use {@link #forNumber(int)} instead. + */ + @java.lang.Deprecated public static Enum1 valueOf(int value) { + return forNumber(value); + } + + /** + * @param value The numeric wire value of the corresponding enum entry. + * @return The enum associated with the given numeric wire value. + */ + public static Enum1 forNumber(int value) { switch (value) { case 1: return VAL1; case 2: return VAL2; @@ -2174,17 +2539,17 @@ public static Enum1 valueOf(int value) { internalGetValueMap() { return internalValueMap; } - private static com.google.protobuf.Internal.EnumLiteMap - internalValueMap = + private static final com.google.protobuf.Internal.EnumLiteMap< + Enum1> internalValueMap = new com.google.protobuf.Internal.EnumLiteMap() { public Enum1 findValueByNumber(int number) { - return Enum1.valueOf(number); + return Enum1.forNumber(number); } }; public final com.google.protobuf.Descriptors.EnumValueDescriptor getValueDescriptor() { - return getDescriptor().getValues().get(index); + return getDescriptor().getValues().get(ordinal()); } public final com.google.protobuf.Descriptors.EnumDescriptor getDescriptorForType() { @@ -2206,11 +2571,9 @@ public static Enum1 valueOf( return VALUES[desc.getIndex()]; } - private final int index; private final int value; - private Enum1(int index, int value) { - this.index = index; + private Enum1(int value) { this.value = value; } @@ -2218,226 +2581,268 @@ private Enum1(int index, int value) { } private int bitField0_; - // optional double doubleType = 1; public static final int DOUBLETYPE_FIELD_NUMBER = 1; private double doubleType_; /** * optional double doubleType = 1; + * @return Whether the doubleType field is set. */ + @java.lang.Override public boolean hasDoubleType() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional double doubleType = 1; + * @return The doubleType. */ + @java.lang.Override public double getDoubleType() { return doubleType_; } - // optional float floatType = 2; public static final int FLOATTYPE_FIELD_NUMBER = 2; private float floatType_; /** * optional float floatType = 2; + * @return Whether the floatType field is set. */ + @java.lang.Override public boolean hasFloatType() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * optional float floatType = 2; + * @return The floatType. */ + @java.lang.Override public float getFloatType() { return floatType_; } - // optional int32 int32Type = 3; public static final int INT32TYPE_FIELD_NUMBER = 3; private int int32Type_; /** * optional int32 int32Type = 3; + * @return Whether the int32Type field is set. */ + @java.lang.Override public boolean hasInt32Type() { - return ((bitField0_ & 0x00000004) == 0x00000004); + return ((bitField0_ & 0x00000004) != 0); } /** * optional int32 int32Type = 3; + * @return The int32Type. */ + @java.lang.Override public int getInt32Type() { return int32Type_; } - // optional int64 int64Type = 4; public static final int INT64TYPE_FIELD_NUMBER = 4; private long int64Type_; /** * optional int64 int64Type = 4; + * @return Whether the int64Type field is set. */ + @java.lang.Override public boolean hasInt64Type() { - return ((bitField0_ & 0x00000008) == 0x00000008); + return ((bitField0_ & 0x00000008) != 0); } /** * optional int64 int64Type = 4; + * @return The int64Type. */ + @java.lang.Override public long getInt64Type() { return int64Type_; } - // optional uint32 uint32Type = 5; public static final int UINT32TYPE_FIELD_NUMBER = 5; private int uint32Type_; /** * optional uint32 uint32Type = 5; + * @return Whether the uint32Type field is set. */ + @java.lang.Override public boolean hasUint32Type() { - return ((bitField0_ & 0x00000010) == 0x00000010); + return ((bitField0_ & 0x00000010) != 0); } /** * optional uint32 uint32Type = 5; + * @return The uint32Type. */ + @java.lang.Override public int getUint32Type() { return uint32Type_; } - // optional uint64 uint64Type = 6; public static final int UINT64TYPE_FIELD_NUMBER = 6; private long uint64Type_; /** * optional uint64 uint64Type = 6; + * @return Whether the uint64Type field is set. */ + @java.lang.Override public boolean hasUint64Type() { - return ((bitField0_ & 0x00000020) == 0x00000020); + return ((bitField0_ & 0x00000020) != 0); } /** * optional uint64 uint64Type = 6; + * @return The uint64Type. */ + @java.lang.Override public long getUint64Type() { return uint64Type_; } - // optional sint32 sint32Type = 7; public static final int SINT32TYPE_FIELD_NUMBER = 7; private int sint32Type_; /** * optional sint32 sint32Type = 7; + * @return Whether the sint32Type field is set. */ + @java.lang.Override public boolean hasSint32Type() { - return ((bitField0_ & 0x00000040) == 0x00000040); + return ((bitField0_ & 0x00000040) != 0); } /** * optional sint32 sint32Type = 7; + * @return The sint32Type. */ + @java.lang.Override public int getSint32Type() { return sint32Type_; } - // optional sint64 sint64Type = 8; public static final int SINT64TYPE_FIELD_NUMBER = 8; private long sint64Type_; /** * optional sint64 sint64Type = 8; + * @return Whether the sint64Type field is set. */ + @java.lang.Override public boolean hasSint64Type() { - return ((bitField0_ & 0x00000080) == 0x00000080); + return ((bitField0_ & 0x00000080) != 0); } /** * optional sint64 sint64Type = 8; + * @return The sint64Type. */ + @java.lang.Override public long getSint64Type() { return sint64Type_; } - // optional fixed32 fixed32Type = 9; public static final int FIXED32TYPE_FIELD_NUMBER = 9; private int fixed32Type_; /** * optional fixed32 fixed32Type = 9; + * @return Whether the fixed32Type field is set. */ + @java.lang.Override public boolean hasFixed32Type() { - return ((bitField0_ & 0x00000100) == 0x00000100); + return ((bitField0_ & 0x00000100) != 0); } /** * optional fixed32 fixed32Type = 9; + * @return The fixed32Type. */ + @java.lang.Override public int getFixed32Type() { return fixed32Type_; } - // optional fixed64 fixed64Type = 10; public static final int FIXED64TYPE_FIELD_NUMBER = 10; private long fixed64Type_; /** * optional fixed64 fixed64Type = 10; + * @return Whether the fixed64Type field is set. */ + @java.lang.Override public boolean hasFixed64Type() { - return ((bitField0_ & 0x00000200) == 0x00000200); + return ((bitField0_ & 0x00000200) != 0); } /** * optional fixed64 fixed64Type = 10; + * @return The fixed64Type. */ + @java.lang.Override public long getFixed64Type() { return fixed64Type_; } - // optional sfixed32 sfixed32Type = 11; public static final int SFIXED32TYPE_FIELD_NUMBER = 11; private int sfixed32Type_; /** * optional sfixed32 sfixed32Type = 11; + * @return Whether the sfixed32Type field is set. */ + @java.lang.Override public boolean hasSfixed32Type() { - return ((bitField0_ & 0x00000400) == 0x00000400); + return ((bitField0_ & 0x00000400) != 0); } /** * optional sfixed32 sfixed32Type = 11; + * @return The sfixed32Type. */ + @java.lang.Override public int getSfixed32Type() { return sfixed32Type_; } - // optional sfixed64 sfixed64Type = 12; public static final int SFIXED64TYPE_FIELD_NUMBER = 12; private long sfixed64Type_; /** * optional sfixed64 sfixed64Type = 12; + * @return Whether the sfixed64Type field is set. */ + @java.lang.Override public boolean hasSfixed64Type() { - return ((bitField0_ & 0x00000800) == 0x00000800); + return ((bitField0_ & 0x00000800) != 0); } /** * optional sfixed64 sfixed64Type = 12; + * @return The sfixed64Type. */ + @java.lang.Override public long getSfixed64Type() { return sfixed64Type_; } - // optional bool boolType = 13; public static final int BOOLTYPE_FIELD_NUMBER = 13; private boolean boolType_; /** * optional bool boolType = 13; + * @return Whether the boolType field is set. */ + @java.lang.Override public boolean hasBoolType() { - return ((bitField0_ & 0x00001000) == 0x00001000); + return ((bitField0_ & 0x00001000) != 0); } /** * optional bool boolType = 13; + * @return The boolType. */ + @java.lang.Override public boolean getBoolType() { return boolType_; } - // optional string stringType = 14; public static final int STRINGTYPE_FIELD_NUMBER = 14; - private java.lang.Object stringType_; + private volatile java.lang.Object stringType_; /** * optional string stringType = 14; + * @return Whether the stringType field is set. */ + @java.lang.Override public boolean hasStringType() { - return ((bitField0_ & 0x00002000) == 0x00002000); + return ((bitField0_ & 0x00002000) != 0); } /** * optional string stringType = 14; + * @return The stringType. */ + @java.lang.Override public java.lang.String getStringType() { java.lang.Object ref = stringType_; if (ref instanceof java.lang.String) { @@ -2454,7 +2859,9 @@ public java.lang.String getStringType() { } /** * optional string stringType = 14; + * @return The bytes for stringType. */ + @java.lang.Override public com.google.protobuf.ByteString getStringTypeBytes() { java.lang.Object ref = stringType_; @@ -2469,34 +2876,38 @@ public java.lang.String getStringType() { } } - // optional bytes bytesType = 15; public static final int BYTESTYPE_FIELD_NUMBER = 15; private com.google.protobuf.ByteString bytesType_; /** * optional bytes bytesType = 15; + * @return Whether the bytesType field is set. */ + @java.lang.Override public boolean hasBytesType() { - return ((bitField0_ & 0x00004000) == 0x00004000); + return ((bitField0_ & 0x00004000) != 0); } /** * optional bytes bytesType = 15; + * @return The bytesType. */ + @java.lang.Override public com.google.protobuf.ByteString getBytesType() { return bytesType_; } - // repeated .MapFieldEntry mapType = 16; public static final int MAPTYPE_FIELD_NUMBER = 16; private java.util.List mapType_; /** * repeated .MapFieldEntry mapType = 16; */ + @java.lang.Override public java.util.List getMapTypeList() { return mapType_; } /** * repeated .MapFieldEntry mapType = 16; */ + @java.lang.Override public java.util.List getMapTypeOrBuilderList() { return mapType_; @@ -2504,87 +2915,100 @@ public java.util.Listrepeated .MapFieldEntry mapType = 16; */ + @java.lang.Override public int getMapTypeCount() { return mapType_.size(); } /** * repeated .MapFieldEntry mapType = 16; */ + @java.lang.Override public org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry getMapType(int index) { return mapType_.get(index); } /** * repeated .MapFieldEntry mapType = 16; */ + @java.lang.Override public org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntryOrBuilder getMapTypeOrBuilder( int index) { return mapType_.get(index); } - // repeated string stringListType = 17; public static final int STRINGLISTTYPE_FIELD_NUMBER = 17; private com.google.protobuf.LazyStringList stringListType_; /** * repeated string stringListType = 17; + * @return A list containing the stringListType. */ - public java.util.List + public com.google.protobuf.ProtocolStringList getStringListTypeList() { return stringListType_; } /** * repeated string stringListType = 17; + * @return The count of stringListType. */ public int getStringListTypeCount() { return stringListType_.size(); } /** * repeated string stringListType = 17; + * @param index The index of the element to return. + * @return The stringListType at the given index. */ public java.lang.String getStringListType(int index) { return stringListType_.get(index); } /** * repeated string stringListType = 17; + * @param index The index of the value to return. + * @return The bytes of the stringListType at the given index. */ public com.google.protobuf.ByteString getStringListTypeBytes(int index) { return stringListType_.getByteString(index); } - // optional .Mesg1 messageType = 18; public static final int MESSAGETYPE_FIELD_NUMBER = 18; private org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1 messageType_; /** * optional .Mesg1 messageType = 18; + * @return Whether the messageType field is set. */ + @java.lang.Override public boolean hasMessageType() { - return ((bitField0_ & 0x00008000) == 0x00008000); + return ((bitField0_ & 0x00008000) != 0); } /** * optional .Mesg1 messageType = 18; + * @return The messageType. */ + @java.lang.Override public org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1 getMessageType() { - return messageType_; + return messageType_ == null ? org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1.getDefaultInstance() : messageType_; } /** * optional .Mesg1 messageType = 18; */ + @java.lang.Override public org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1OrBuilder getMessageTypeOrBuilder() { - return messageType_; + return messageType_ == null ? org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1.getDefaultInstance() : messageType_; } - // repeated .Mesg1 messageListType = 19; public static final int MESSAGELISTTYPE_FIELD_NUMBER = 19; private java.util.List messageListType_; /** * repeated .Mesg1 messageListType = 19; */ + @java.lang.Override public java.util.List getMessageListTypeList() { return messageListType_; } /** * repeated .Mesg1 messageListType = 19; */ + @java.lang.Override public java.util.List getMessageListTypeOrBuilderList() { return messageListType_; @@ -2592,199 +3016,184 @@ public java.util.List /** * repeated .Mesg1 messageListType = 19; */ + @java.lang.Override public int getMessageListTypeCount() { return messageListType_.size(); } /** * repeated .Mesg1 messageListType = 19; */ + @java.lang.Override public org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1 getMessageListType(int index) { return messageListType_.get(index); } /** * repeated .Mesg1 messageListType = 19; */ + @java.lang.Override public org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1OrBuilder getMessageListTypeOrBuilder( int index) { return messageListType_.get(index); } - // optional .AllTypes.Enum1 enumType = 20; public static final int ENUMTYPE_FIELD_NUMBER = 20; - private org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes.Enum1 enumType_; + private int enumType_; /** * optional .AllTypes.Enum1 enumType = 20; + * @return Whether the enumType field is set. */ - public boolean hasEnumType() { - return ((bitField0_ & 0x00010000) == 0x00010000); + @java.lang.Override public boolean hasEnumType() { + return ((bitField0_ & 0x00010000) != 0); } /** * optional .AllTypes.Enum1 enumType = 20; + * @return The enumType. */ - public org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes.Enum1 getEnumType() { - return enumType_; - } - - private void initFields() { - doubleType_ = 0D; - floatType_ = 0F; - int32Type_ = 0; - int64Type_ = 0L; - uint32Type_ = 0; - uint64Type_ = 0L; - sint32Type_ = 0; - sint64Type_ = 0L; - fixed32Type_ = 0; - fixed64Type_ = 0L; - sfixed32Type_ = 0; - sfixed64Type_ = 0L; - boolType_ = false; - stringType_ = ""; - bytesType_ = com.google.protobuf.ByteString.EMPTY; - mapType_ = java.util.Collections.emptyList(); - stringListType_ = com.google.protobuf.LazyStringArrayList.EMPTY; - messageType_ = org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1.getDefaultInstance(); - messageListType_ = java.util.Collections.emptyList(); - enumType_ = org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes.Enum1.VAL1; + @java.lang.Override public org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes.Enum1 getEnumType() { + @SuppressWarnings("deprecation") + org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes.Enum1 result = org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes.Enum1.valueOf(enumType_); + return result == null ? org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes.Enum1.VAL1 : result; } + private byte memoizedIsInitialized = -1; + @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; - if (isInitialized != -1) return isInitialized == 1; + if (isInitialized == 1) return true; + if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } + @java.lang.Override public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { - getSerializedSize(); - if (((bitField0_ & 0x00000001) == 0x00000001)) { + if (((bitField0_ & 0x00000001) != 0)) { output.writeDouble(1, doubleType_); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { + if (((bitField0_ & 0x00000002) != 0)) { output.writeFloat(2, floatType_); } - if (((bitField0_ & 0x00000004) == 0x00000004)) { + if (((bitField0_ & 0x00000004) != 0)) { output.writeInt32(3, int32Type_); } - if (((bitField0_ & 0x00000008) == 0x00000008)) { + if (((bitField0_ & 0x00000008) != 0)) { output.writeInt64(4, int64Type_); } - if (((bitField0_ & 0x00000010) == 0x00000010)) { + if (((bitField0_ & 0x00000010) != 0)) { output.writeUInt32(5, uint32Type_); } - if (((bitField0_ & 0x00000020) == 0x00000020)) { + if (((bitField0_ & 0x00000020) != 0)) { output.writeUInt64(6, uint64Type_); } - if (((bitField0_ & 0x00000040) == 0x00000040)) { + if (((bitField0_ & 0x00000040) != 0)) { output.writeSInt32(7, sint32Type_); } - if (((bitField0_ & 0x00000080) == 0x00000080)) { + if (((bitField0_ & 0x00000080) != 0)) { output.writeSInt64(8, sint64Type_); } - if (((bitField0_ & 0x00000100) == 0x00000100)) { + if (((bitField0_ & 0x00000100) != 0)) { output.writeFixed32(9, fixed32Type_); } - if (((bitField0_ & 0x00000200) == 0x00000200)) { + if (((bitField0_ & 0x00000200) != 0)) { output.writeFixed64(10, fixed64Type_); } - if (((bitField0_ & 0x00000400) == 0x00000400)) { + if (((bitField0_ & 0x00000400) != 0)) { output.writeSFixed32(11, sfixed32Type_); } - if (((bitField0_ & 0x00000800) == 0x00000800)) { + if (((bitField0_ & 0x00000800) != 0)) { output.writeSFixed64(12, sfixed64Type_); } - if (((bitField0_ & 0x00001000) == 0x00001000)) { + if (((bitField0_ & 0x00001000) != 0)) { output.writeBool(13, boolType_); } - if (((bitField0_ & 0x00002000) == 0x00002000)) { - output.writeBytes(14, getStringTypeBytes()); + if (((bitField0_ & 0x00002000) != 0)) { + com.google.protobuf.GeneratedMessageV3.writeString(output, 14, stringType_); } - if (((bitField0_ & 0x00004000) == 0x00004000)) { + if (((bitField0_ & 0x00004000) != 0)) { output.writeBytes(15, bytesType_); } for (int i = 0; i < mapType_.size(); i++) { output.writeMessage(16, mapType_.get(i)); } for (int i = 0; i < stringListType_.size(); i++) { - output.writeBytes(17, stringListType_.getByteString(i)); + com.google.protobuf.GeneratedMessageV3.writeString(output, 17, stringListType_.getRaw(i)); } - if (((bitField0_ & 0x00008000) == 0x00008000)) { - output.writeMessage(18, messageType_); + if (((bitField0_ & 0x00008000) != 0)) { + output.writeMessage(18, getMessageType()); } for (int i = 0; i < messageListType_.size(); i++) { output.writeMessage(19, messageListType_.get(i)); } - if (((bitField0_ & 0x00010000) == 0x00010000)) { - output.writeEnum(20, enumType_.getNumber()); + if (((bitField0_ & 0x00010000) != 0)) { + output.writeEnum(20, enumType_); } - getUnknownFields().writeTo(output); + unknownFields.writeTo(output); } - private int memoizedSerializedSize = -1; + @java.lang.Override public int getSerializedSize() { - int size = memoizedSerializedSize; + int size = memoizedSize; if (size != -1) return size; size = 0; - if (((bitField0_ & 0x00000001) == 0x00000001)) { + if (((bitField0_ & 0x00000001) != 0)) { size += com.google.protobuf.CodedOutputStream .computeDoubleSize(1, doubleType_); } - if (((bitField0_ & 0x00000002) == 0x00000002)) { + if (((bitField0_ & 0x00000002) != 0)) { size += com.google.protobuf.CodedOutputStream .computeFloatSize(2, floatType_); } - if (((bitField0_ & 0x00000004) == 0x00000004)) { + if (((bitField0_ & 0x00000004) != 0)) { size += com.google.protobuf.CodedOutputStream .computeInt32Size(3, int32Type_); } - if (((bitField0_ & 0x00000008) == 0x00000008)) { + if (((bitField0_ & 0x00000008) != 0)) { size += com.google.protobuf.CodedOutputStream .computeInt64Size(4, int64Type_); } - if (((bitField0_ & 0x00000010) == 0x00000010)) { + if (((bitField0_ & 0x00000010) != 0)) { size += com.google.protobuf.CodedOutputStream .computeUInt32Size(5, uint32Type_); } - if (((bitField0_ & 0x00000020) == 0x00000020)) { + if (((bitField0_ & 0x00000020) != 0)) { size += com.google.protobuf.CodedOutputStream .computeUInt64Size(6, uint64Type_); } - if (((bitField0_ & 0x00000040) == 0x00000040)) { + if (((bitField0_ & 0x00000040) != 0)) { size += com.google.protobuf.CodedOutputStream .computeSInt32Size(7, sint32Type_); } - if (((bitField0_ & 0x00000080) == 0x00000080)) { + if (((bitField0_ & 0x00000080) != 0)) { size += com.google.protobuf.CodedOutputStream .computeSInt64Size(8, sint64Type_); } - if (((bitField0_ & 0x00000100) == 0x00000100)) { + if (((bitField0_ & 0x00000100) != 0)) { size += com.google.protobuf.CodedOutputStream .computeFixed32Size(9, fixed32Type_); } - if (((bitField0_ & 0x00000200) == 0x00000200)) { + if (((bitField0_ & 0x00000200) != 0)) { size += com.google.protobuf.CodedOutputStream .computeFixed64Size(10, fixed64Type_); } - if (((bitField0_ & 0x00000400) == 0x00000400)) { + if (((bitField0_ & 0x00000400) != 0)) { size += com.google.protobuf.CodedOutputStream .computeSFixed32Size(11, sfixed32Type_); } - if (((bitField0_ & 0x00000800) == 0x00000800)) { + if (((bitField0_ & 0x00000800) != 0)) { size += com.google.protobuf.CodedOutputStream .computeSFixed64Size(12, sfixed64Type_); } - if (((bitField0_ & 0x00001000) == 0x00001000)) { + if (((bitField0_ & 0x00001000) != 0)) { size += com.google.protobuf.CodedOutputStream .computeBoolSize(13, boolType_); } - if (((bitField0_ & 0x00002000) == 0x00002000)) { - size += com.google.protobuf.CodedOutputStream - .computeBytesSize(14, getStringTypeBytes()); + if (((bitField0_ & 0x00002000) != 0)) { + size += com.google.protobuf.GeneratedMessageV3.computeStringSize(14, stringType_); } - if (((bitField0_ & 0x00004000) == 0x00004000)) { + if (((bitField0_ & 0x00004000) != 0)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(15, bytesType_); } @@ -2795,36 +3204,245 @@ public int getSerializedSize() { { int dataSize = 0; for (int i = 0; i < stringListType_.size(); i++) { - dataSize += com.google.protobuf.CodedOutputStream - .computeBytesSizeNoTag(stringListType_.getByteString(i)); + dataSize += computeStringSizeNoTag(stringListType_.getRaw(i)); } size += dataSize; size += 2 * getStringListTypeList().size(); } - if (((bitField0_ & 0x00008000) == 0x00008000)) { + if (((bitField0_ & 0x00008000) != 0)) { size += com.google.protobuf.CodedOutputStream - .computeMessageSize(18, messageType_); + .computeMessageSize(18, getMessageType()); } for (int i = 0; i < messageListType_.size(); i++) { size += com.google.protobuf.CodedOutputStream .computeMessageSize(19, messageListType_.get(i)); } - if (((bitField0_ & 0x00010000) == 0x00010000)) { + if (((bitField0_ & 0x00010000) != 0)) { size += com.google.protobuf.CodedOutputStream - .computeEnumSize(20, enumType_.getNumber()); + .computeEnumSize(20, enumType_); } - size += getUnknownFields().getSerializedSize(); - memoizedSerializedSize = size; + size += unknownFields.getSerializedSize(); + memoizedSize = size; return size; } - private static final long serialVersionUID = 0L; @java.lang.Override - protected java.lang.Object writeReplace() - throws java.io.ObjectStreamException { - return super.writeReplace(); + public boolean equals(final java.lang.Object obj) { + if (obj == this) { + return true; + } + if (!(obj instanceof org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes)) { + return super.equals(obj); + } + org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes other = (org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes) obj; + + if (hasDoubleType() != other.hasDoubleType()) return false; + if (hasDoubleType()) { + if (java.lang.Double.doubleToLongBits(getDoubleType()) + != java.lang.Double.doubleToLongBits( + other.getDoubleType())) return false; + } + if (hasFloatType() != other.hasFloatType()) return false; + if (hasFloatType()) { + if (java.lang.Float.floatToIntBits(getFloatType()) + != java.lang.Float.floatToIntBits( + other.getFloatType())) return false; + } + if (hasInt32Type() != other.hasInt32Type()) return false; + if (hasInt32Type()) { + if (getInt32Type() + != other.getInt32Type()) return false; + } + if (hasInt64Type() != other.hasInt64Type()) return false; + if (hasInt64Type()) { + if (getInt64Type() + != other.getInt64Type()) return false; + } + if (hasUint32Type() != other.hasUint32Type()) return false; + if (hasUint32Type()) { + if (getUint32Type() + != other.getUint32Type()) return false; + } + if (hasUint64Type() != other.hasUint64Type()) return false; + if (hasUint64Type()) { + if (getUint64Type() + != other.getUint64Type()) return false; + } + if (hasSint32Type() != other.hasSint32Type()) return false; + if (hasSint32Type()) { + if (getSint32Type() + != other.getSint32Type()) return false; + } + if (hasSint64Type() != other.hasSint64Type()) return false; + if (hasSint64Type()) { + if (getSint64Type() + != other.getSint64Type()) return false; + } + if (hasFixed32Type() != other.hasFixed32Type()) return false; + if (hasFixed32Type()) { + if (getFixed32Type() + != other.getFixed32Type()) return false; + } + if (hasFixed64Type() != other.hasFixed64Type()) return false; + if (hasFixed64Type()) { + if (getFixed64Type() + != other.getFixed64Type()) return false; + } + if (hasSfixed32Type() != other.hasSfixed32Type()) return false; + if (hasSfixed32Type()) { + if (getSfixed32Type() + != other.getSfixed32Type()) return false; + } + if (hasSfixed64Type() != other.hasSfixed64Type()) return false; + if (hasSfixed64Type()) { + if (getSfixed64Type() + != other.getSfixed64Type()) return false; + } + if (hasBoolType() != other.hasBoolType()) return false; + if (hasBoolType()) { + if (getBoolType() + != other.getBoolType()) return false; + } + if (hasStringType() != other.hasStringType()) return false; + if (hasStringType()) { + if (!getStringType() + .equals(other.getStringType())) return false; + } + if (hasBytesType() != other.hasBytesType()) return false; + if (hasBytesType()) { + if (!getBytesType() + .equals(other.getBytesType())) return false; + } + if (!getMapTypeList() + .equals(other.getMapTypeList())) return false; + if (!getStringListTypeList() + .equals(other.getStringListTypeList())) return false; + if (hasMessageType() != other.hasMessageType()) return false; + if (hasMessageType()) { + if (!getMessageType() + .equals(other.getMessageType())) return false; + } + if (!getMessageListTypeList() + .equals(other.getMessageListTypeList())) return false; + if (hasEnumType() != other.hasEnumType()) return false; + if (hasEnumType()) { + if (enumType_ != other.enumType_) return false; + } + if (!unknownFields.equals(other.unknownFields)) return false; + return true; } + @java.lang.Override + public int hashCode() { + if (memoizedHashCode != 0) { + return memoizedHashCode; + } + int hash = 41; + hash = (19 * hash) + getDescriptor().hashCode(); + if (hasDoubleType()) { + hash = (37 * hash) + DOUBLETYPE_FIELD_NUMBER; + hash = (53 * hash) + com.google.protobuf.Internal.hashLong( + java.lang.Double.doubleToLongBits(getDoubleType())); + } + if (hasFloatType()) { + hash = (37 * hash) + FLOATTYPE_FIELD_NUMBER; + hash = (53 * hash) + java.lang.Float.floatToIntBits( + getFloatType()); + } + if (hasInt32Type()) { + hash = (37 * hash) + INT32TYPE_FIELD_NUMBER; + hash = (53 * hash) + getInt32Type(); + } + if (hasInt64Type()) { + hash = (37 * hash) + INT64TYPE_FIELD_NUMBER; + hash = (53 * hash) + com.google.protobuf.Internal.hashLong( + getInt64Type()); + } + if (hasUint32Type()) { + hash = (37 * hash) + UINT32TYPE_FIELD_NUMBER; + hash = (53 * hash) + getUint32Type(); + } + if (hasUint64Type()) { + hash = (37 * hash) + UINT64TYPE_FIELD_NUMBER; + hash = (53 * hash) + com.google.protobuf.Internal.hashLong( + getUint64Type()); + } + if (hasSint32Type()) { + hash = (37 * hash) + SINT32TYPE_FIELD_NUMBER; + hash = (53 * hash) + getSint32Type(); + } + if (hasSint64Type()) { + hash = (37 * hash) + SINT64TYPE_FIELD_NUMBER; + hash = (53 * hash) + com.google.protobuf.Internal.hashLong( + getSint64Type()); + } + if (hasFixed32Type()) { + hash = (37 * hash) + FIXED32TYPE_FIELD_NUMBER; + hash = (53 * hash) + getFixed32Type(); + } + if (hasFixed64Type()) { + hash = (37 * hash) + FIXED64TYPE_FIELD_NUMBER; + hash = (53 * hash) + com.google.protobuf.Internal.hashLong( + getFixed64Type()); + } + if (hasSfixed32Type()) { + hash = (37 * hash) + SFIXED32TYPE_FIELD_NUMBER; + hash = (53 * hash) + getSfixed32Type(); + } + if (hasSfixed64Type()) { + hash = (37 * hash) + SFIXED64TYPE_FIELD_NUMBER; + hash = (53 * hash) + com.google.protobuf.Internal.hashLong( + getSfixed64Type()); + } + if (hasBoolType()) { + hash = (37 * hash) + BOOLTYPE_FIELD_NUMBER; + hash = (53 * hash) + com.google.protobuf.Internal.hashBoolean( + getBoolType()); + } + if (hasStringType()) { + hash = (37 * hash) + STRINGTYPE_FIELD_NUMBER; + hash = (53 * hash) + getStringType().hashCode(); + } + if (hasBytesType()) { + hash = (37 * hash) + BYTESTYPE_FIELD_NUMBER; + hash = (53 * hash) + getBytesType().hashCode(); + } + if (getMapTypeCount() > 0) { + hash = (37 * hash) + MAPTYPE_FIELD_NUMBER; + hash = (53 * hash) + getMapTypeList().hashCode(); + } + if (getStringListTypeCount() > 0) { + hash = (37 * hash) + STRINGLISTTYPE_FIELD_NUMBER; + hash = (53 * hash) + getStringListTypeList().hashCode(); + } + if (hasMessageType()) { + hash = (37 * hash) + MESSAGETYPE_FIELD_NUMBER; + hash = (53 * hash) + getMessageType().hashCode(); + } + if (getMessageListTypeCount() > 0) { + hash = (37 * hash) + MESSAGELISTTYPE_FIELD_NUMBER; + hash = (53 * hash) + getMessageListTypeList().hashCode(); + } + if (hasEnumType()) { + hash = (37 * hash) + ENUMTYPE_FIELD_NUMBER; + hash = (53 * hash) + enumType_; + } + hash = (29 * hash) + unknownFields.hashCode(); + memoizedHashCode = hash; + return hash; + } + + public static org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes parseFrom( + java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data); + } + public static org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes parseFrom( + java.nio.ByteBuffer data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return PARSER.parseFrom(data, extensionRegistry); + } public static org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { @@ -2848,46 +3466,59 @@ public static org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes parseF } public static org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes parseFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseDelimitedFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { - return PARSER.parseFrom(input); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { - return PARSER.parseFrom(input, extensionRegistry); + return com.google.protobuf.GeneratedMessageV3 + .parseWithIOException(PARSER, input, extensionRegistry); } - public static Builder newBuilder() { return Builder.create(); } + @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } + public static Builder newBuilder() { + return DEFAULT_INSTANCE.toBuilder(); + } public static Builder newBuilder(org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes prototype) { - return newBuilder().mergeFrom(prototype); + return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); + } + @java.lang.Override + public Builder toBuilder() { + return this == DEFAULT_INSTANCE + ? new Builder() : new Builder().mergeFrom(this); } - public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } @@ -2895,14 +3526,16 @@ protected Builder newBuilderForType( * Protobuf type {@code AllTypes} */ public static final class Builder extends - com.google.protobuf.GeneratedMessage.Builder - implements org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypesOrBuilder { + com.google.protobuf.GeneratedMessageV3.Builder implements + // @@protoc_insertion_point(builder_implements:AllTypes) + org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypesOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.internal_static_AllTypes_descriptor; } - protected com.google.protobuf.GeneratedMessage.FieldAccessorTable + @java.lang.Override + protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.internal_static_AllTypes_fieldAccessorTable .ensureFieldAccessorsInitialized( @@ -2915,21 +3548,19 @@ private Builder() { } private Builder( - com.google.protobuf.GeneratedMessage.BuilderParent parent) { + com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { - if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { + if (com.google.protobuf.GeneratedMessageV3 + .alwaysUseFieldBuilders) { getMapTypeFieldBuilder(); getMessageTypeFieldBuilder(); getMessageListTypeFieldBuilder(); } } - private static Builder create() { - return new Builder(); - } - + @java.lang.Override public Builder clear() { super.clear(); doubleType_ = 0D; @@ -2971,7 +3602,7 @@ public Builder clear() { stringListType_ = com.google.protobuf.LazyStringArrayList.EMPTY; bitField0_ = (bitField0_ & ~0x00010000); if (messageTypeBuilder_ == null) { - messageType_ = org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1.getDefaultInstance(); + messageType_ = null; } else { messageTypeBuilder_.clear(); } @@ -2982,24 +3613,23 @@ public Builder clear() { } else { messageListTypeBuilder_.clear(); } - enumType_ = org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes.Enum1.VAL1; + enumType_ = 1; bitField0_ = (bitField0_ & ~0x00080000); return this; } - public Builder clone() { - return create().mergeFrom(buildPartial()); - } - + @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.internal_static_AllTypes_descriptor; } + @java.lang.Override public org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes getDefaultInstanceForType() { return org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes.getDefaultInstance(); } + @java.lang.Override public org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes build() { org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes result = buildPartial(); if (!result.isInitialized()) { @@ -3008,72 +3638,73 @@ public org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes build() { return result; } + @java.lang.Override public org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes buildPartial() { org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes result = new org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; - if (((from_bitField0_ & 0x00000001) == 0x00000001)) { + if (((from_bitField0_ & 0x00000001) != 0)) { + result.doubleType_ = doubleType_; to_bitField0_ |= 0x00000001; } - result.doubleType_ = doubleType_; - if (((from_bitField0_ & 0x00000002) == 0x00000002)) { + if (((from_bitField0_ & 0x00000002) != 0)) { + result.floatType_ = floatType_; to_bitField0_ |= 0x00000002; } - result.floatType_ = floatType_; - if (((from_bitField0_ & 0x00000004) == 0x00000004)) { + if (((from_bitField0_ & 0x00000004) != 0)) { + result.int32Type_ = int32Type_; to_bitField0_ |= 0x00000004; } - result.int32Type_ = int32Type_; - if (((from_bitField0_ & 0x00000008) == 0x00000008)) { + if (((from_bitField0_ & 0x00000008) != 0)) { + result.int64Type_ = int64Type_; to_bitField0_ |= 0x00000008; } - result.int64Type_ = int64Type_; - if (((from_bitField0_ & 0x00000010) == 0x00000010)) { + if (((from_bitField0_ & 0x00000010) != 0)) { + result.uint32Type_ = uint32Type_; to_bitField0_ |= 0x00000010; } - result.uint32Type_ = uint32Type_; - if (((from_bitField0_ & 0x00000020) == 0x00000020)) { + if (((from_bitField0_ & 0x00000020) != 0)) { + result.uint64Type_ = uint64Type_; to_bitField0_ |= 0x00000020; } - result.uint64Type_ = uint64Type_; - if (((from_bitField0_ & 0x00000040) == 0x00000040)) { + if (((from_bitField0_ & 0x00000040) != 0)) { + result.sint32Type_ = sint32Type_; to_bitField0_ |= 0x00000040; } - result.sint32Type_ = sint32Type_; - if (((from_bitField0_ & 0x00000080) == 0x00000080)) { + if (((from_bitField0_ & 0x00000080) != 0)) { + result.sint64Type_ = sint64Type_; to_bitField0_ |= 0x00000080; } - result.sint64Type_ = sint64Type_; - if (((from_bitField0_ & 0x00000100) == 0x00000100)) { + if (((from_bitField0_ & 0x00000100) != 0)) { + result.fixed32Type_ = fixed32Type_; to_bitField0_ |= 0x00000100; } - result.fixed32Type_ = fixed32Type_; - if (((from_bitField0_ & 0x00000200) == 0x00000200)) { + if (((from_bitField0_ & 0x00000200) != 0)) { + result.fixed64Type_ = fixed64Type_; to_bitField0_ |= 0x00000200; } - result.fixed64Type_ = fixed64Type_; - if (((from_bitField0_ & 0x00000400) == 0x00000400)) { + if (((from_bitField0_ & 0x00000400) != 0)) { + result.sfixed32Type_ = sfixed32Type_; to_bitField0_ |= 0x00000400; } - result.sfixed32Type_ = sfixed32Type_; - if (((from_bitField0_ & 0x00000800) == 0x00000800)) { + if (((from_bitField0_ & 0x00000800) != 0)) { + result.sfixed64Type_ = sfixed64Type_; to_bitField0_ |= 0x00000800; } - result.sfixed64Type_ = sfixed64Type_; - if (((from_bitField0_ & 0x00001000) == 0x00001000)) { + if (((from_bitField0_ & 0x00001000) != 0)) { + result.boolType_ = boolType_; to_bitField0_ |= 0x00001000; } - result.boolType_ = boolType_; - if (((from_bitField0_ & 0x00002000) == 0x00002000)) { + if (((from_bitField0_ & 0x00002000) != 0)) { to_bitField0_ |= 0x00002000; } result.stringType_ = stringType_; - if (((from_bitField0_ & 0x00004000) == 0x00004000)) { + if (((from_bitField0_ & 0x00004000) != 0)) { to_bitField0_ |= 0x00004000; } result.bytesType_ = bytesType_; if (mapTypeBuilder_ == null) { - if (((bitField0_ & 0x00008000) == 0x00008000)) { + if (((bitField0_ & 0x00008000) != 0)) { mapType_ = java.util.Collections.unmodifiableList(mapType_); bitField0_ = (bitField0_ & ~0x00008000); } @@ -3081,22 +3712,21 @@ public org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes buildPartial( } else { result.mapType_ = mapTypeBuilder_.build(); } - if (((bitField0_ & 0x00010000) == 0x00010000)) { - stringListType_ = new com.google.protobuf.UnmodifiableLazyStringList( - stringListType_); + if (((bitField0_ & 0x00010000) != 0)) { + stringListType_ = stringListType_.getUnmodifiableView(); bitField0_ = (bitField0_ & ~0x00010000); } result.stringListType_ = stringListType_; - if (((from_bitField0_ & 0x00020000) == 0x00020000)) { + if (((from_bitField0_ & 0x00020000) != 0)) { + if (messageTypeBuilder_ == null) { + result.messageType_ = messageType_; + } else { + result.messageType_ = messageTypeBuilder_.build(); + } to_bitField0_ |= 0x00008000; } - if (messageTypeBuilder_ == null) { - result.messageType_ = messageType_; - } else { - result.messageType_ = messageTypeBuilder_.build(); - } if (messageListTypeBuilder_ == null) { - if (((bitField0_ & 0x00040000) == 0x00040000)) { + if (((bitField0_ & 0x00040000) != 0)) { messageListType_ = java.util.Collections.unmodifiableList(messageListType_); bitField0_ = (bitField0_ & ~0x00040000); } @@ -3104,7 +3734,7 @@ public org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes buildPartial( } else { result.messageListType_ = messageListTypeBuilder_.build(); } - if (((from_bitField0_ & 0x00080000) == 0x00080000)) { + if (((from_bitField0_ & 0x00080000) != 0)) { to_bitField0_ |= 0x00010000; } result.enumType_ = enumType_; @@ -3113,6 +3743,39 @@ public org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes buildPartial( return result; } + @java.lang.Override + public Builder clone() { + return super.clone(); + } + @java.lang.Override + public Builder setField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.setField(field, value); + } + @java.lang.Override + public Builder clearField( + com.google.protobuf.Descriptors.FieldDescriptor field) { + return super.clearField(field); + } + @java.lang.Override + public Builder clearOneof( + com.google.protobuf.Descriptors.OneofDescriptor oneof) { + return super.clearOneof(oneof); + } + @java.lang.Override + public Builder setRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + int index, java.lang.Object value) { + return super.setRepeatedField(field, index, value); + } + @java.lang.Override + public Builder addRepeatedField( + com.google.protobuf.Descriptors.FieldDescriptor field, + java.lang.Object value) { + return super.addRepeatedField(field, value); + } + @java.lang.Override public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes) { return mergeFrom((org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes)other); @@ -3190,7 +3853,7 @@ public Builder mergeFrom(org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllT mapType_ = other.mapType_; bitField0_ = (bitField0_ & ~0x00008000); mapTypeBuilder_ = - com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? + com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getMapTypeFieldBuilder() : null; } else { mapTypeBuilder_.addAllMessages(other.mapType_); @@ -3229,7 +3892,7 @@ public Builder mergeFrom(org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllT messageListType_ = other.messageListType_; bitField0_ = (bitField0_ & ~0x00040000); messageListTypeBuilder_ = - com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? + com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getMessageListTypeFieldBuilder() : null; } else { messageListTypeBuilder_.addAllMessages(other.messageListType_); @@ -3239,14 +3902,17 @@ public Builder mergeFrom(org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllT if (other.hasEnumType()) { setEnumType(other.getEnumType()); } - this.mergeUnknownFields(other.getUnknownFields()); + this.mergeUnknownFields(other.unknownFields); + onChanged(); return this; } + @java.lang.Override public final boolean isInitialized() { return true; } + @java.lang.Override public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) @@ -3256,7 +3922,7 @@ public Builder mergeFrom( parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes) e.getUnfinishedMessage(); - throw e; + throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); @@ -3266,22 +3932,27 @@ public Builder mergeFrom( } private int bitField0_; - // optional double doubleType = 1; private double doubleType_ ; /** * optional double doubleType = 1; + * @return Whether the doubleType field is set. */ + @java.lang.Override public boolean hasDoubleType() { - return ((bitField0_ & 0x00000001) == 0x00000001); + return ((bitField0_ & 0x00000001) != 0); } /** * optional double doubleType = 1; + * @return The doubleType. */ + @java.lang.Override public double getDoubleType() { return doubleType_; } /** * optional double doubleType = 1; + * @param value The doubleType to set. + * @return This builder for chaining. */ public Builder setDoubleType(double value) { bitField0_ |= 0x00000001; @@ -3291,6 +3962,7 @@ public Builder setDoubleType(double value) { } /** * optional double doubleType = 1; + * @return This builder for chaining. */ public Builder clearDoubleType() { bitField0_ = (bitField0_ & ~0x00000001); @@ -3299,22 +3971,27 @@ public Builder clearDoubleType() { return this; } - // optional float floatType = 2; private float floatType_ ; /** * optional float floatType = 2; + * @return Whether the floatType field is set. */ + @java.lang.Override public boolean hasFloatType() { - return ((bitField0_ & 0x00000002) == 0x00000002); + return ((bitField0_ & 0x00000002) != 0); } /** * optional float floatType = 2; + * @return The floatType. */ + @java.lang.Override public float getFloatType() { return floatType_; } /** * optional float floatType = 2; + * @param value The floatType to set. + * @return This builder for chaining. */ public Builder setFloatType(float value) { bitField0_ |= 0x00000002; @@ -3324,6 +4001,7 @@ public Builder setFloatType(float value) { } /** * optional float floatType = 2; + * @return This builder for chaining. */ public Builder clearFloatType() { bitField0_ = (bitField0_ & ~0x00000002); @@ -3332,22 +4010,27 @@ public Builder clearFloatType() { return this; } - // optional int32 int32Type = 3; private int int32Type_ ; /** * optional int32 int32Type = 3; + * @return Whether the int32Type field is set. */ + @java.lang.Override public boolean hasInt32Type() { - return ((bitField0_ & 0x00000004) == 0x00000004); + return ((bitField0_ & 0x00000004) != 0); } /** * optional int32 int32Type = 3; + * @return The int32Type. */ + @java.lang.Override public int getInt32Type() { return int32Type_; } /** * optional int32 int32Type = 3; + * @param value The int32Type to set. + * @return This builder for chaining. */ public Builder setInt32Type(int value) { bitField0_ |= 0x00000004; @@ -3357,6 +4040,7 @@ public Builder setInt32Type(int value) { } /** * optional int32 int32Type = 3; + * @return This builder for chaining. */ public Builder clearInt32Type() { bitField0_ = (bitField0_ & ~0x00000004); @@ -3365,22 +4049,27 @@ public Builder clearInt32Type() { return this; } - // optional int64 int64Type = 4; private long int64Type_ ; /** * optional int64 int64Type = 4; + * @return Whether the int64Type field is set. */ + @java.lang.Override public boolean hasInt64Type() { - return ((bitField0_ & 0x00000008) == 0x00000008); + return ((bitField0_ & 0x00000008) != 0); } /** * optional int64 int64Type = 4; + * @return The int64Type. */ + @java.lang.Override public long getInt64Type() { return int64Type_; } /** * optional int64 int64Type = 4; + * @param value The int64Type to set. + * @return This builder for chaining. */ public Builder setInt64Type(long value) { bitField0_ |= 0x00000008; @@ -3390,6 +4079,7 @@ public Builder setInt64Type(long value) { } /** * optional int64 int64Type = 4; + * @return This builder for chaining. */ public Builder clearInt64Type() { bitField0_ = (bitField0_ & ~0x00000008); @@ -3398,22 +4088,27 @@ public Builder clearInt64Type() { return this; } - // optional uint32 uint32Type = 5; private int uint32Type_ ; /** * optional uint32 uint32Type = 5; + * @return Whether the uint32Type field is set. */ + @java.lang.Override public boolean hasUint32Type() { - return ((bitField0_ & 0x00000010) == 0x00000010); + return ((bitField0_ & 0x00000010) != 0); } /** * optional uint32 uint32Type = 5; + * @return The uint32Type. */ + @java.lang.Override public int getUint32Type() { return uint32Type_; } /** * optional uint32 uint32Type = 5; + * @param value The uint32Type to set. + * @return This builder for chaining. */ public Builder setUint32Type(int value) { bitField0_ |= 0x00000010; @@ -3423,6 +4118,7 @@ public Builder setUint32Type(int value) { } /** * optional uint32 uint32Type = 5; + * @return This builder for chaining. */ public Builder clearUint32Type() { bitField0_ = (bitField0_ & ~0x00000010); @@ -3431,22 +4127,27 @@ public Builder clearUint32Type() { return this; } - // optional uint64 uint64Type = 6; private long uint64Type_ ; /** * optional uint64 uint64Type = 6; + * @return Whether the uint64Type field is set. */ + @java.lang.Override public boolean hasUint64Type() { - return ((bitField0_ & 0x00000020) == 0x00000020); + return ((bitField0_ & 0x00000020) != 0); } /** * optional uint64 uint64Type = 6; + * @return The uint64Type. */ + @java.lang.Override public long getUint64Type() { return uint64Type_; } /** * optional uint64 uint64Type = 6; + * @param value The uint64Type to set. + * @return This builder for chaining. */ public Builder setUint64Type(long value) { bitField0_ |= 0x00000020; @@ -3456,6 +4157,7 @@ public Builder setUint64Type(long value) { } /** * optional uint64 uint64Type = 6; + * @return This builder for chaining. */ public Builder clearUint64Type() { bitField0_ = (bitField0_ & ~0x00000020); @@ -3464,22 +4166,27 @@ public Builder clearUint64Type() { return this; } - // optional sint32 sint32Type = 7; private int sint32Type_ ; /** * optional sint32 sint32Type = 7; + * @return Whether the sint32Type field is set. */ + @java.lang.Override public boolean hasSint32Type() { - return ((bitField0_ & 0x00000040) == 0x00000040); + return ((bitField0_ & 0x00000040) != 0); } /** * optional sint32 sint32Type = 7; + * @return The sint32Type. */ + @java.lang.Override public int getSint32Type() { return sint32Type_; } /** * optional sint32 sint32Type = 7; + * @param value The sint32Type to set. + * @return This builder for chaining. */ public Builder setSint32Type(int value) { bitField0_ |= 0x00000040; @@ -3489,6 +4196,7 @@ public Builder setSint32Type(int value) { } /** * optional sint32 sint32Type = 7; + * @return This builder for chaining. */ public Builder clearSint32Type() { bitField0_ = (bitField0_ & ~0x00000040); @@ -3497,22 +4205,27 @@ public Builder clearSint32Type() { return this; } - // optional sint64 sint64Type = 8; private long sint64Type_ ; /** * optional sint64 sint64Type = 8; + * @return Whether the sint64Type field is set. */ + @java.lang.Override public boolean hasSint64Type() { - return ((bitField0_ & 0x00000080) == 0x00000080); + return ((bitField0_ & 0x00000080) != 0); } /** * optional sint64 sint64Type = 8; + * @return The sint64Type. */ + @java.lang.Override public long getSint64Type() { return sint64Type_; } /** * optional sint64 sint64Type = 8; + * @param value The sint64Type to set. + * @return This builder for chaining. */ public Builder setSint64Type(long value) { bitField0_ |= 0x00000080; @@ -3522,6 +4235,7 @@ public Builder setSint64Type(long value) { } /** * optional sint64 sint64Type = 8; + * @return This builder for chaining. */ public Builder clearSint64Type() { bitField0_ = (bitField0_ & ~0x00000080); @@ -3530,22 +4244,27 @@ public Builder clearSint64Type() { return this; } - // optional fixed32 fixed32Type = 9; private int fixed32Type_ ; /** * optional fixed32 fixed32Type = 9; + * @return Whether the fixed32Type field is set. */ + @java.lang.Override public boolean hasFixed32Type() { - return ((bitField0_ & 0x00000100) == 0x00000100); + return ((bitField0_ & 0x00000100) != 0); } /** * optional fixed32 fixed32Type = 9; + * @return The fixed32Type. */ + @java.lang.Override public int getFixed32Type() { return fixed32Type_; } /** * optional fixed32 fixed32Type = 9; + * @param value The fixed32Type to set. + * @return This builder for chaining. */ public Builder setFixed32Type(int value) { bitField0_ |= 0x00000100; @@ -3555,6 +4274,7 @@ public Builder setFixed32Type(int value) { } /** * optional fixed32 fixed32Type = 9; + * @return This builder for chaining. */ public Builder clearFixed32Type() { bitField0_ = (bitField0_ & ~0x00000100); @@ -3563,22 +4283,27 @@ public Builder clearFixed32Type() { return this; } - // optional fixed64 fixed64Type = 10; private long fixed64Type_ ; /** * optional fixed64 fixed64Type = 10; + * @return Whether the fixed64Type field is set. */ + @java.lang.Override public boolean hasFixed64Type() { - return ((bitField0_ & 0x00000200) == 0x00000200); + return ((bitField0_ & 0x00000200) != 0); } /** * optional fixed64 fixed64Type = 10; + * @return The fixed64Type. */ + @java.lang.Override public long getFixed64Type() { return fixed64Type_; } /** * optional fixed64 fixed64Type = 10; + * @param value The fixed64Type to set. + * @return This builder for chaining. */ public Builder setFixed64Type(long value) { bitField0_ |= 0x00000200; @@ -3588,6 +4313,7 @@ public Builder setFixed64Type(long value) { } /** * optional fixed64 fixed64Type = 10; + * @return This builder for chaining. */ public Builder clearFixed64Type() { bitField0_ = (bitField0_ & ~0x00000200); @@ -3596,22 +4322,27 @@ public Builder clearFixed64Type() { return this; } - // optional sfixed32 sfixed32Type = 11; private int sfixed32Type_ ; /** * optional sfixed32 sfixed32Type = 11; + * @return Whether the sfixed32Type field is set. */ + @java.lang.Override public boolean hasSfixed32Type() { - return ((bitField0_ & 0x00000400) == 0x00000400); + return ((bitField0_ & 0x00000400) != 0); } /** * optional sfixed32 sfixed32Type = 11; + * @return The sfixed32Type. */ + @java.lang.Override public int getSfixed32Type() { return sfixed32Type_; } /** * optional sfixed32 sfixed32Type = 11; + * @param value The sfixed32Type to set. + * @return This builder for chaining. */ public Builder setSfixed32Type(int value) { bitField0_ |= 0x00000400; @@ -3621,6 +4352,7 @@ public Builder setSfixed32Type(int value) { } /** * optional sfixed32 sfixed32Type = 11; + * @return This builder for chaining. */ public Builder clearSfixed32Type() { bitField0_ = (bitField0_ & ~0x00000400); @@ -3629,22 +4361,27 @@ public Builder clearSfixed32Type() { return this; } - // optional sfixed64 sfixed64Type = 12; private long sfixed64Type_ ; /** * optional sfixed64 sfixed64Type = 12; + * @return Whether the sfixed64Type field is set. */ + @java.lang.Override public boolean hasSfixed64Type() { - return ((bitField0_ & 0x00000800) == 0x00000800); + return ((bitField0_ & 0x00000800) != 0); } /** * optional sfixed64 sfixed64Type = 12; + * @return The sfixed64Type. */ + @java.lang.Override public long getSfixed64Type() { return sfixed64Type_; } /** * optional sfixed64 sfixed64Type = 12; + * @param value The sfixed64Type to set. + * @return This builder for chaining. */ public Builder setSfixed64Type(long value) { bitField0_ |= 0x00000800; @@ -3654,6 +4391,7 @@ public Builder setSfixed64Type(long value) { } /** * optional sfixed64 sfixed64Type = 12; + * @return This builder for chaining. */ public Builder clearSfixed64Type() { bitField0_ = (bitField0_ & ~0x00000800); @@ -3662,22 +4400,27 @@ public Builder clearSfixed64Type() { return this; } - // optional bool boolType = 13; private boolean boolType_ ; /** * optional bool boolType = 13; + * @return Whether the boolType field is set. */ + @java.lang.Override public boolean hasBoolType() { - return ((bitField0_ & 0x00001000) == 0x00001000); + return ((bitField0_ & 0x00001000) != 0); } /** * optional bool boolType = 13; + * @return The boolType. */ + @java.lang.Override public boolean getBoolType() { return boolType_; } /** * optional bool boolType = 13; + * @param value The boolType to set. + * @return This builder for chaining. */ public Builder setBoolType(boolean value) { bitField0_ |= 0x00001000; @@ -3687,6 +4430,7 @@ public Builder setBoolType(boolean value) { } /** * optional bool boolType = 13; + * @return This builder for chaining. */ public Builder clearBoolType() { bitField0_ = (bitField0_ & ~0x00001000); @@ -3695,23 +4439,27 @@ public Builder clearBoolType() { return this; } - // optional string stringType = 14; private java.lang.Object stringType_ = ""; /** * optional string stringType = 14; + * @return Whether the stringType field is set. */ public boolean hasStringType() { - return ((bitField0_ & 0x00002000) == 0x00002000); + return ((bitField0_ & 0x00002000) != 0); } /** * optional string stringType = 14; + * @return The stringType. */ public java.lang.String getStringType() { java.lang.Object ref = stringType_; if (!(ref instanceof java.lang.String)) { - java.lang.String s = ((com.google.protobuf.ByteString) ref) - .toStringUtf8(); - stringType_ = s; + com.google.protobuf.ByteString bs = + (com.google.protobuf.ByteString) ref; + java.lang.String s = bs.toStringUtf8(); + if (bs.isValidUtf8()) { + stringType_ = s; + } return s; } else { return (java.lang.String) ref; @@ -3719,6 +4467,7 @@ public java.lang.String getStringType() { } /** * optional string stringType = 14; + * @return The bytes for stringType. */ public com.google.protobuf.ByteString getStringTypeBytes() { @@ -3735,6 +4484,8 @@ public java.lang.String getStringType() { } /** * optional string stringType = 14; + * @param value The stringType to set. + * @return This builder for chaining. */ public Builder setStringType( java.lang.String value) { @@ -3748,6 +4499,7 @@ public Builder setStringType( } /** * optional string stringType = 14; + * @return This builder for chaining. */ public Builder clearStringType() { bitField0_ = (bitField0_ & ~0x00002000); @@ -3757,6 +4509,8 @@ public Builder clearStringType() { } /** * optional string stringType = 14; + * @param value The bytes for stringType to set. + * @return This builder for chaining. */ public Builder setStringTypeBytes( com.google.protobuf.ByteString value) { @@ -3769,22 +4523,27 @@ public Builder setStringTypeBytes( return this; } - // optional bytes bytesType = 15; private com.google.protobuf.ByteString bytesType_ = com.google.protobuf.ByteString.EMPTY; /** * optional bytes bytesType = 15; + * @return Whether the bytesType field is set. */ + @java.lang.Override public boolean hasBytesType() { - return ((bitField0_ & 0x00004000) == 0x00004000); + return ((bitField0_ & 0x00004000) != 0); } /** * optional bytes bytesType = 15; + * @return The bytesType. */ + @java.lang.Override public com.google.protobuf.ByteString getBytesType() { return bytesType_; } /** * optional bytes bytesType = 15; + * @param value The bytesType to set. + * @return This builder for chaining. */ public Builder setBytesType(com.google.protobuf.ByteString value) { if (value == null) { @@ -3797,6 +4556,7 @@ public Builder setBytesType(com.google.protobuf.ByteString value) { } /** * optional bytes bytesType = 15; + * @return This builder for chaining. */ public Builder clearBytesType() { bitField0_ = (bitField0_ & ~0x00004000); @@ -3805,17 +4565,16 @@ public Builder clearBytesType() { return this; } - // repeated .MapFieldEntry mapType = 16; private java.util.List mapType_ = java.util.Collections.emptyList(); private void ensureMapTypeIsMutable() { - if (!((bitField0_ & 0x00008000) == 0x00008000)) { + if (!((bitField0_ & 0x00008000) != 0)) { mapType_ = new java.util.ArrayList(mapType_); bitField0_ |= 0x00008000; } } - private com.google.protobuf.RepeatedFieldBuilder< + private com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry, org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry.Builder, org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntryOrBuilder> mapTypeBuilder_; /** @@ -3947,7 +4706,8 @@ public Builder addAllMapType( java.lang.Iterable values) { if (mapTypeBuilder_ == null) { ensureMapTypeIsMutable(); - super.addAll(values, mapType_); + com.google.protobuf.AbstractMessageLite.Builder.addAll( + values, mapType_); onChanged(); } else { mapTypeBuilder_.addAllMessages(values); @@ -4030,14 +4790,14 @@ public org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry.Builder getMapTypeBuilderList() { return getMapTypeFieldBuilder().getBuilderList(); } - private com.google.protobuf.RepeatedFieldBuilder< + private com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry, org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry.Builder, org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntryOrBuilder> getMapTypeFieldBuilder() { if (mapTypeBuilder_ == null) { - mapTypeBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< + mapTypeBuilder_ = new com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry, org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry.Builder, org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntryOrBuilder>( mapType_, - ((bitField0_ & 0x00008000) == 0x00008000), + ((bitField0_ & 0x00008000) != 0), getParentForChildren(), isClean()); mapType_ = null; @@ -4045,35 +4805,40 @@ public org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.MapFieldEntry.Builder return mapTypeBuilder_; } - // repeated string stringListType = 17; private com.google.protobuf.LazyStringList stringListType_ = com.google.protobuf.LazyStringArrayList.EMPTY; private void ensureStringListTypeIsMutable() { - if (!((bitField0_ & 0x00010000) == 0x00010000)) { + if (!((bitField0_ & 0x00010000) != 0)) { stringListType_ = new com.google.protobuf.LazyStringArrayList(stringListType_); bitField0_ |= 0x00010000; } } /** * repeated string stringListType = 17; + * @return A list containing the stringListType. */ - public java.util.List + public com.google.protobuf.ProtocolStringList getStringListTypeList() { - return java.util.Collections.unmodifiableList(stringListType_); + return stringListType_.getUnmodifiableView(); } /** * repeated string stringListType = 17; + * @return The count of stringListType. */ public int getStringListTypeCount() { return stringListType_.size(); } /** * repeated string stringListType = 17; + * @param index The index of the element to return. + * @return The stringListType at the given index. */ public java.lang.String getStringListType(int index) { return stringListType_.get(index); } /** * repeated string stringListType = 17; + * @param index The index of the value to return. + * @return The bytes of the stringListType at the given index. */ public com.google.protobuf.ByteString getStringListTypeBytes(int index) { @@ -4081,6 +4846,9 @@ public java.lang.String getStringListType(int index) { } /** * repeated string stringListType = 17; + * @param index The index to set the value at. + * @param value The stringListType to set. + * @return This builder for chaining. */ public Builder setStringListType( int index, java.lang.String value) { @@ -4094,6 +4862,8 @@ public Builder setStringListType( } /** * repeated string stringListType = 17; + * @param value The stringListType to add. + * @return This builder for chaining. */ public Builder addStringListType( java.lang.String value) { @@ -4107,16 +4877,20 @@ public Builder addStringListType( } /** * repeated string stringListType = 17; + * @param values The stringListType to add. + * @return This builder for chaining. */ public Builder addAllStringListType( java.lang.Iterable values) { ensureStringListTypeIsMutable(); - super.addAll(values, stringListType_); + com.google.protobuf.AbstractMessageLite.Builder.addAll( + values, stringListType_); onChanged(); return this; } /** * repeated string stringListType = 17; + * @return This builder for chaining. */ public Builder clearStringListType() { stringListType_ = com.google.protobuf.LazyStringArrayList.EMPTY; @@ -4126,6 +4900,8 @@ public Builder clearStringListType() { } /** * repeated string stringListType = 17; + * @param value The bytes of the stringListType to add. + * @return This builder for chaining. */ public Builder addStringListTypeBytes( com.google.protobuf.ByteString value) { @@ -4138,22 +4914,23 @@ public Builder addStringListTypeBytes( return this; } - // optional .Mesg1 messageType = 18; - private org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1 messageType_ = org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1.getDefaultInstance(); - private com.google.protobuf.SingleFieldBuilder< + private org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1 messageType_; + private com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1, org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1.Builder, org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1OrBuilder> messageTypeBuilder_; /** * optional .Mesg1 messageType = 18; + * @return Whether the messageType field is set. */ public boolean hasMessageType() { - return ((bitField0_ & 0x00020000) == 0x00020000); + return ((bitField0_ & 0x00020000) != 0); } /** * optional .Mesg1 messageType = 18; + * @return The messageType. */ public org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1 getMessageType() { if (messageTypeBuilder_ == null) { - return messageType_; + return messageType_ == null ? org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1.getDefaultInstance() : messageType_; } else { return messageTypeBuilder_.getMessage(); } @@ -4193,7 +4970,8 @@ public Builder setMessageType( */ public Builder mergeMessageType(org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1 value) { if (messageTypeBuilder_ == null) { - if (((bitField0_ & 0x00020000) == 0x00020000) && + if (((bitField0_ & 0x00020000) != 0) && + messageType_ != null && messageType_ != org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1.getDefaultInstance()) { messageType_ = org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1.newBuilder(messageType_).mergeFrom(value).buildPartial(); @@ -4212,7 +4990,7 @@ public Builder mergeMessageType(org.apache.hadoop.hive.ql.io.protobuf.SampleProt */ public Builder clearMessageType() { if (messageTypeBuilder_ == null) { - messageType_ = org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1.getDefaultInstance(); + messageType_ = null; onChanged(); } else { messageTypeBuilder_.clear(); @@ -4235,19 +5013,20 @@ public org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1OrBuilder getMess if (messageTypeBuilder_ != null) { return messageTypeBuilder_.getMessageOrBuilder(); } else { - return messageType_; + return messageType_ == null ? + org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1.getDefaultInstance() : messageType_; } } /** * optional .Mesg1 messageType = 18; */ - private com.google.protobuf.SingleFieldBuilder< + private com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1, org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1.Builder, org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1OrBuilder> getMessageTypeFieldBuilder() { if (messageTypeBuilder_ == null) { - messageTypeBuilder_ = new com.google.protobuf.SingleFieldBuilder< + messageTypeBuilder_ = new com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1, org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1.Builder, org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1OrBuilder>( - messageType_, + getMessageType(), getParentForChildren(), isClean()); messageType_ = null; @@ -4255,17 +5034,16 @@ public org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1OrBuilder getMess return messageTypeBuilder_; } - // repeated .Mesg1 messageListType = 19; private java.util.List messageListType_ = java.util.Collections.emptyList(); private void ensureMessageListTypeIsMutable() { - if (!((bitField0_ & 0x00040000) == 0x00040000)) { + if (!((bitField0_ & 0x00040000) != 0)) { messageListType_ = new java.util.ArrayList(messageListType_); bitField0_ |= 0x00040000; } } - private com.google.protobuf.RepeatedFieldBuilder< + private com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1, org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1.Builder, org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1OrBuilder> messageListTypeBuilder_; /** @@ -4397,7 +5175,8 @@ public Builder addAllMessageListType( java.lang.Iterable values) { if (messageListTypeBuilder_ == null) { ensureMessageListTypeIsMutable(); - super.addAll(values, messageListType_); + com.google.protobuf.AbstractMessageLite.Builder.addAll( + values, messageListType_); onChanged(); } else { messageListTypeBuilder_.addAllMessages(values); @@ -4480,14 +5259,14 @@ public org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1.Builder addMessa getMessageListTypeBuilderList() { return getMessageListTypeFieldBuilder().getBuilderList(); } - private com.google.protobuf.RepeatedFieldBuilder< + private com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1, org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1.Builder, org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1OrBuilder> getMessageListTypeFieldBuilder() { if (messageListTypeBuilder_ == null) { - messageListTypeBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< + messageListTypeBuilder_ = new com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1, org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1.Builder, org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1OrBuilder>( messageListType_, - ((bitField0_ & 0x00040000) == 0x00040000), + ((bitField0_ & 0x00040000) != 0), getParentForChildren(), isClean()); messageListType_ = null; @@ -4495,74 +5274,122 @@ public org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.Mesg1.Builder addMessa return messageListTypeBuilder_; } - // optional .AllTypes.Enum1 enumType = 20; - private org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes.Enum1 enumType_ = org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes.Enum1.VAL1; + private int enumType_ = 1; /** * optional .AllTypes.Enum1 enumType = 20; + * @return Whether the enumType field is set. */ - public boolean hasEnumType() { - return ((bitField0_ & 0x00080000) == 0x00080000); + @java.lang.Override public boolean hasEnumType() { + return ((bitField0_ & 0x00080000) != 0); } /** * optional .AllTypes.Enum1 enumType = 20; + * @return The enumType. */ + @java.lang.Override public org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes.Enum1 getEnumType() { - return enumType_; + @SuppressWarnings("deprecation") + org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes.Enum1 result = org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes.Enum1.valueOf(enumType_); + return result == null ? org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes.Enum1.VAL1 : result; } /** * optional .AllTypes.Enum1 enumType = 20; + * @param value The enumType to set. + * @return This builder for chaining. */ public Builder setEnumType(org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes.Enum1 value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00080000; - enumType_ = value; + enumType_ = value.getNumber(); onChanged(); return this; } /** * optional .AllTypes.Enum1 enumType = 20; + * @return This builder for chaining. */ public Builder clearEnumType() { bitField0_ = (bitField0_ & ~0x00080000); - enumType_ = org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes.Enum1.VAL1; + enumType_ = 1; onChanged(); return this; } + @java.lang.Override + public final Builder setUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.setUnknownFields(unknownFields); + } + + @java.lang.Override + public final Builder mergeUnknownFields( + final com.google.protobuf.UnknownFieldSet unknownFields) { + return super.mergeUnknownFields(unknownFields); + } + // @@protoc_insertion_point(builder_scope:AllTypes) } + // @@protoc_insertion_point(class_scope:AllTypes) + private static final org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes DEFAULT_INSTANCE; static { - defaultInstance = new AllTypes(true); - defaultInstance.initFields(); + DEFAULT_INSTANCE = new org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes(); + } + + public static org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes getDefaultInstance() { + return DEFAULT_INSTANCE; + } + + @java.lang.Deprecated public static final com.google.protobuf.Parser + PARSER = new com.google.protobuf.AbstractParser() { + @java.lang.Override + public AllTypes parsePartialFrom( + com.google.protobuf.CodedInputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return new AllTypes(input, extensionRegistry); + } + }; + + public static com.google.protobuf.Parser parser() { + return PARSER; + } + + @java.lang.Override + public com.google.protobuf.Parser getParserForType() { + return PARSER; + } + + @java.lang.Override + public org.apache.hadoop.hive.ql.io.protobuf.SampleProtos.AllTypes getDefaultInstanceForType() { + return DEFAULT_INSTANCE; } - // @@protoc_insertion_point(class_scope:AllTypes) } - private static com.google.protobuf.Descriptors.Descriptor + private static final com.google.protobuf.Descriptors.Descriptor internal_static_MapFieldEntry_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + private static final + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_MapFieldEntry_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static final com.google.protobuf.Descriptors.Descriptor internal_static_Mesg1_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + private static final + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_Mesg1_fieldAccessorTable; - private static com.google.protobuf.Descriptors.Descriptor + private static final com.google.protobuf.Descriptors.Descriptor internal_static_AllTypes_descriptor; - private static - com.google.protobuf.GeneratedMessage.FieldAccessorTable + private static final + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_AllTypes_fieldAccessorTable; public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { return descriptor; } - private static com.google.protobuf.Descriptors.FileDescriptor + private static com.google.protobuf.Descriptors.FileDescriptor descriptor; static { java.lang.String[] descriptorData = { @@ -4575,7 +5402,7 @@ public Builder clearEnumType() { "pe\030\004 \001(\003\022\022\n\nuint32Type\030\005 \001(\r\022\022\n\nuint64Ty" + "pe\030\006 \001(\004\022\022\n\nsint32Type\030\007 \001(\021\022\022\n\nsint64Ty" + "pe\030\010 \001(\022\022\023\n\013fixed32Type\030\t \001(\007\022\023\n\013fixed64" + - "Type\030\n \001(\006\022\024\n\014sfixed32Type\030\013 \001(\017\022\024\n\014sfix", + "Type\030\n \001(\006\022\024\n\014sfixed32Type\030\013 \001(\017\022\024\n\014sfix" + "ed64Type\030\014 \001(\020\022\020\n\010boolType\030\r \001(\010\022\022\n\nstri" + "ngType\030\016 \001(\t\022\021\n\tbytesType\030\017 \001(\014\022\037\n\007mapTy" + "pe\030\020 \003(\0132\016.MapFieldEntry\022\026\n\016stringListTy" + @@ -4585,36 +5412,28 @@ public Builder clearEnumType() { "L1\020\001\022\010\n\004VAL2\020\002B5\n%org.apache.hadoop.hive" + ".ql.io.protobufB\014SampleProtos" }; - com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = - new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { - public com.google.protobuf.ExtensionRegistry assignDescriptors( - com.google.protobuf.Descriptors.FileDescriptor root) { - descriptor = root; - internal_static_MapFieldEntry_descriptor = - getDescriptor().getMessageTypes().get(0); - internal_static_MapFieldEntry_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_MapFieldEntry_descriptor, - new java.lang.String[] { "Key", "Value", }); - internal_static_Mesg1_descriptor = - getDescriptor().getMessageTypes().get(1); - internal_static_Mesg1_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_Mesg1_descriptor, - new java.lang.String[] { "AnotherMap", "NoMap", "IntList", }); - internal_static_AllTypes_descriptor = - getDescriptor().getMessageTypes().get(2); - internal_static_AllTypes_fieldAccessorTable = new - com.google.protobuf.GeneratedMessage.FieldAccessorTable( - internal_static_AllTypes_descriptor, - new java.lang.String[] { "DoubleType", "FloatType", "Int32Type", "Int64Type", "Uint32Type", "Uint64Type", "Sint32Type", "Sint64Type", "Fixed32Type", "Fixed64Type", "Sfixed32Type", "Sfixed64Type", "BoolType", "StringType", "BytesType", "MapType", "StringListType", "MessageType", "MessageListType", "EnumType", }); - return null; - } - }; - com.google.protobuf.Descriptors.FileDescriptor + descriptor = com.google.protobuf.Descriptors.FileDescriptor .internalBuildGeneratedFileFrom(descriptorData, new com.google.protobuf.Descriptors.FileDescriptor[] { - }, assigner); + }); + internal_static_MapFieldEntry_descriptor = + getDescriptor().getMessageTypes().get(0); + internal_static_MapFieldEntry_fieldAccessorTable = new + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_MapFieldEntry_descriptor, + new java.lang.String[] { "Key", "Value", }); + internal_static_Mesg1_descriptor = + getDescriptor().getMessageTypes().get(1); + internal_static_Mesg1_fieldAccessorTable = new + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_Mesg1_descriptor, + new java.lang.String[] { "AnotherMap", "NoMap", "IntList", }); + internal_static_AllTypes_descriptor = + getDescriptor().getMessageTypes().get(2); + internal_static_AllTypes_fieldAccessorTable = new + com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_AllTypes_descriptor, + new java.lang.String[] { "DoubleType", "FloatType", "Int32Type", "Int64Type", "Uint32Type", "Uint64Type", "Sint32Type", "Sint64Type", "Fixed32Type", "Fixed64Type", "Sfixed32Type", "Sfixed64Type", "BoolType", "StringType", "BytesType", "MapType", "StringListType", "MessageType", "MessageListType", "EnumType", }); } // @@protoc_insertion_point(outer_class_scope) diff --git a/ql/src/java/org/apache/tez/dag/history/logging/proto/ProtoMessageWritable.java b/ql/src/java/org/apache/tez/dag/history/logging/proto/ProtoMessageWritable.java index 7a08e202d3ae..2fb94194e788 100644 --- a/ql/src/java/org/apache/tez/dag/history/logging/proto/ProtoMessageWritable.java +++ b/ql/src/java/org/apache/tez/dag/history/logging/proto/ProtoMessageWritable.java @@ -25,6 +25,7 @@ import java.io.InputStream; import java.io.OutputStream; +import com.google.protobuf.ExtensionRegistryLite; import org.apache.hadoop.io.Writable; import com.google.protobuf.CodedInputStream; @@ -40,6 +41,8 @@ public class ProtoMessageWritable implements Writable { private DataInputStream din; private CodedInputStream cin; + private final ExtensionRegistryLite extensionRegistryLite = ExtensionRegistryLite.newInstance(); + ProtoMessageWritable(Parser parser) { this.parser = parser; } @@ -96,6 +99,6 @@ public void readFields(DataInput in) throws IOException { cin.setSizeLimit(Integer.MAX_VALUE); } din.in = in; - message = cin.readMessage(parser, null); + message = cin.readMessage(parser, extensionRegistryLite); } } diff --git a/ql/src/protobuf/java/HiveEvents.proto b/ql/src/protobuf/java/HiveEvents.proto index eab0cc91c16f..ca8540105b9f 100644 --- a/ql/src/protobuf/java/HiveEvents.proto +++ b/ql/src/protobuf/java/HiveEvents.proto @@ -15,6 +15,7 @@ * See the License for the specific language governing permissions and * limitations under the License. */ +syntax = "proto2"; option java_package = "org.apache.hadoop.hive.ql.hooks.proto"; option java_outer_classname = "HiveHookEvents"; diff --git a/standalone-metastore/metastore-common/pom.xml b/standalone-metastore/metastore-common/pom.xml index c6c670a7dcea..319eaf64f38f 100644 --- a/standalone-metastore/metastore-common/pom.xml +++ b/standalone-metastore/metastore-common/pom.xml @@ -449,7 +449,7 @@ com.github.os72 protoc-jar-maven-plugin - 3.5.1.1 + 3.11.4 generate-sources @@ -457,7 +457,7 @@ run - ${protobuf.group}:protoc:${protobuf-exc.version} + com.google.protobuf:protoc:${protobuf.version} none ${basedir}/src/main/protobuf/org/apache/hadoop/hive/metastore diff --git a/standalone-metastore/metastore-common/src/main/protobuf/org/apache/hadoop/hive/metastore/metastore.proto b/standalone-metastore/metastore-common/src/main/protobuf/org/apache/hadoop/hive/metastore/metastore.proto index 29b99b4f669f..dd9d2e49524f 100644 --- a/standalone-metastore/metastore-common/src/main/protobuf/org/apache/hadoop/hive/metastore/metastore.proto +++ b/standalone-metastore/metastore-common/src/main/protobuf/org/apache/hadoop/hive/metastore/metastore.proto @@ -15,13 +15,13 @@ * See the License for the specific language governing permissions and * limitations under the License. */ +syntax = "proto3"; package org.apache.hadoop.hive.metastore; - message SplitInfo { - required int64 offset = 1; - required int64 length = 2; - required int32 index = 3; + int64 offset = 1; + int64 length = 2; + int32 index = 3; } message SplitInfos { diff --git a/standalone-metastore/pom.xml b/standalone-metastore/pom.xml index 1599907c63c7..283cb91996bb 100644 --- a/standalone-metastore/pom.xml +++ b/standalone-metastore/pom.xml @@ -87,10 +87,7 @@ 2.17.1 3.3.3 1.6.9 - - com.google.protobuf - 2.6.1 - 2.6.1 + 3.21.1 1.9.0 2.14.6 4.0.4 @@ -526,20 +523,6 @@ - - - aarch64 - - com.github.os72 - 2.6.1-build3 - - - - linux - aarch64 - - - javadoc