HdfsServerFederationProtos.java
// Generated by the protocol buffer compiler. DO NOT EDIT!
// source: FederationProtocol.proto
// Protobuf Java Version: 3.25.5
package org.apache.hadoop.hdfs.federation.protocol.proto;
public final class HdfsServerFederationProtos {
private HdfsServerFederationProtos() {}
public static void registerAllExtensions(
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite registry) {
}
public static void registerAllExtensions(
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistry registry) {
registerAllExtensions(
(org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite) registry);
}
public interface NamenodeMembershipStatsRecordProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.NamenodeMembershipStatsRecordProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* <code>optional uint64 totalSpace = 1;</code>
* @return Whether the totalSpace field is set.
*/
boolean hasTotalSpace();
/**
* <code>optional uint64 totalSpace = 1;</code>
* @return The totalSpace.
*/
long getTotalSpace();
/**
* <code>optional uint64 availableSpace = 2;</code>
* @return Whether the availableSpace field is set.
*/
boolean hasAvailableSpace();
/**
* <code>optional uint64 availableSpace = 2;</code>
* @return The availableSpace.
*/
long getAvailableSpace();
/**
* <code>optional uint64 providedSpace = 3;</code>
* @return Whether the providedSpace field is set.
*/
boolean hasProvidedSpace();
/**
* <code>optional uint64 providedSpace = 3;</code>
* @return The providedSpace.
*/
long getProvidedSpace();
/**
* <code>optional uint64 numOfFiles = 10;</code>
* @return Whether the numOfFiles field is set.
*/
boolean hasNumOfFiles();
/**
* <code>optional uint64 numOfFiles = 10;</code>
* @return The numOfFiles.
*/
long getNumOfFiles();
/**
* <code>optional uint64 numOfBlocks = 11;</code>
* @return Whether the numOfBlocks field is set.
*/
boolean hasNumOfBlocks();
/**
* <code>optional uint64 numOfBlocks = 11;</code>
* @return The numOfBlocks.
*/
long getNumOfBlocks();
/**
* <code>optional uint64 numOfBlocksMissing = 12;</code>
* @return Whether the numOfBlocksMissing field is set.
*/
boolean hasNumOfBlocksMissing();
/**
* <code>optional uint64 numOfBlocksMissing = 12;</code>
* @return The numOfBlocksMissing.
*/
long getNumOfBlocksMissing();
/**
* <code>optional uint64 numOfBlocksPendingReplication = 13;</code>
* @return Whether the numOfBlocksPendingReplication field is set.
*/
boolean hasNumOfBlocksPendingReplication();
/**
* <code>optional uint64 numOfBlocksPendingReplication = 13;</code>
* @return The numOfBlocksPendingReplication.
*/
long getNumOfBlocksPendingReplication();
/**
* <code>optional uint64 numOfBlocksUnderReplicated = 14;</code>
* @return Whether the numOfBlocksUnderReplicated field is set.
*/
boolean hasNumOfBlocksUnderReplicated();
/**
* <code>optional uint64 numOfBlocksUnderReplicated = 14;</code>
* @return The numOfBlocksUnderReplicated.
*/
long getNumOfBlocksUnderReplicated();
/**
* <code>optional uint64 numOfBlocksPendingDeletion = 15;</code>
* @return Whether the numOfBlocksPendingDeletion field is set.
*/
boolean hasNumOfBlocksPendingDeletion();
/**
* <code>optional uint64 numOfBlocksPendingDeletion = 15;</code>
* @return The numOfBlocksPendingDeletion.
*/
long getNumOfBlocksPendingDeletion();
/**
* <code>optional uint32 numOfActiveDatanodes = 20;</code>
* @return Whether the numOfActiveDatanodes field is set.
*/
boolean hasNumOfActiveDatanodes();
/**
* <code>optional uint32 numOfActiveDatanodes = 20;</code>
* @return The numOfActiveDatanodes.
*/
int getNumOfActiveDatanodes();
/**
* <code>optional uint32 numOfDeadDatanodes = 21;</code>
* @return Whether the numOfDeadDatanodes field is set.
*/
boolean hasNumOfDeadDatanodes();
/**
* <code>optional uint32 numOfDeadDatanodes = 21;</code>
* @return The numOfDeadDatanodes.
*/
int getNumOfDeadDatanodes();
/**
* <code>optional uint32 numOfDecommissioningDatanodes = 22;</code>
* @return Whether the numOfDecommissioningDatanodes field is set.
*/
boolean hasNumOfDecommissioningDatanodes();
/**
* <code>optional uint32 numOfDecommissioningDatanodes = 22;</code>
* @return The numOfDecommissioningDatanodes.
*/
int getNumOfDecommissioningDatanodes();
/**
* <code>optional uint32 numOfDecomActiveDatanodes = 23;</code>
* @return Whether the numOfDecomActiveDatanodes field is set.
*/
boolean hasNumOfDecomActiveDatanodes();
/**
* <code>optional uint32 numOfDecomActiveDatanodes = 23;</code>
* @return The numOfDecomActiveDatanodes.
*/
int getNumOfDecomActiveDatanodes();
/**
* <code>optional uint32 numOfDecomDeadDatanodes = 24;</code>
* @return Whether the numOfDecomDeadDatanodes field is set.
*/
boolean hasNumOfDecomDeadDatanodes();
/**
* <code>optional uint32 numOfDecomDeadDatanodes = 24;</code>
* @return The numOfDecomDeadDatanodes.
*/
int getNumOfDecomDeadDatanodes();
/**
* <code>optional uint32 numOfStaleDatanodes = 25;</code>
* @return Whether the numOfStaleDatanodes field is set.
*/
boolean hasNumOfStaleDatanodes();
/**
* <code>optional uint32 numOfStaleDatanodes = 25;</code>
* @return The numOfStaleDatanodes.
*/
int getNumOfStaleDatanodes();
/**
* <code>optional uint32 numOfInMaintenanceLiveDataNodes = 26;</code>
* @return Whether the numOfInMaintenanceLiveDataNodes field is set.
*/
boolean hasNumOfInMaintenanceLiveDataNodes();
/**
* <code>optional uint32 numOfInMaintenanceLiveDataNodes = 26;</code>
* @return The numOfInMaintenanceLiveDataNodes.
*/
int getNumOfInMaintenanceLiveDataNodes();
/**
* <code>optional uint32 numOfInMaintenanceDeadDataNodes = 27;</code>
* @return Whether the numOfInMaintenanceDeadDataNodes field is set.
*/
boolean hasNumOfInMaintenanceDeadDataNodes();
/**
* <code>optional uint32 numOfInMaintenanceDeadDataNodes = 27;</code>
* @return The numOfInMaintenanceDeadDataNodes.
*/
int getNumOfInMaintenanceDeadDataNodes();
/**
* <code>optional uint32 numOfEnteringMaintenanceDataNodes = 28;</code>
* @return Whether the numOfEnteringMaintenanceDataNodes field is set.
*/
boolean hasNumOfEnteringMaintenanceDataNodes();
/**
* <code>optional uint32 numOfEnteringMaintenanceDataNodes = 28;</code>
* @return The numOfEnteringMaintenanceDataNodes.
*/
int getNumOfEnteringMaintenanceDataNodes();
/**
* <code>optional uint32 corruptFilesCount = 29;</code>
* @return Whether the corruptFilesCount field is set.
*/
boolean hasCorruptFilesCount();
/**
* <code>optional uint32 corruptFilesCount = 29;</code>
* @return The corruptFilesCount.
*/
int getCorruptFilesCount();
/**
* <code>optional uint64 scheduledReplicationBlocks = 30;</code>
* @return Whether the scheduledReplicationBlocks field is set.
*/
boolean hasScheduledReplicationBlocks();
/**
* <code>optional uint64 scheduledReplicationBlocks = 30;</code>
* @return The scheduledReplicationBlocks.
*/
long getScheduledReplicationBlocks();
/**
* <code>optional uint64 numberOfMissingBlocksWithReplicationFactorOne = 31;</code>
* @return Whether the numberOfMissingBlocksWithReplicationFactorOne field is set.
*/
boolean hasNumberOfMissingBlocksWithReplicationFactorOne();
/**
* <code>optional uint64 numberOfMissingBlocksWithReplicationFactorOne = 31;</code>
* @return The numberOfMissingBlocksWithReplicationFactorOne.
*/
long getNumberOfMissingBlocksWithReplicationFactorOne();
/**
* <code>optional uint64 highestPriorityLowRedundancyReplicatedBlocks = 32;</code>
* @return Whether the highestPriorityLowRedundancyReplicatedBlocks field is set.
*/
boolean hasHighestPriorityLowRedundancyReplicatedBlocks();
/**
* <code>optional uint64 highestPriorityLowRedundancyReplicatedBlocks = 32;</code>
* @return The highestPriorityLowRedundancyReplicatedBlocks.
*/
long getHighestPriorityLowRedundancyReplicatedBlocks();
/**
* <code>optional uint64 HighestPriorityLowRedundancyECBlocks = 33;</code>
* @return Whether the highestPriorityLowRedundancyECBlocks field is set.
*/
boolean hasHighestPriorityLowRedundancyECBlocks();
/**
* <code>optional uint64 HighestPriorityLowRedundancyECBlocks = 33;</code>
* @return The highestPriorityLowRedundancyECBlocks.
*/
long getHighestPriorityLowRedundancyECBlocks();
/**
* <code>optional uint32 pendingSPSPaths = 34;</code>
* @return Whether the pendingSPSPaths field is set.
*/
boolean hasPendingSPSPaths();
/**
* <code>optional uint32 pendingSPSPaths = 34;</code>
* @return The pendingSPSPaths.
*/
int getPendingSPSPaths();
/**
* <code>optional uint64 badlyDistributedBlocks = 35;</code>
* @return Whether the badlyDistributedBlocks field is set.
*/
boolean hasBadlyDistributedBlocks();
/**
* <code>optional uint64 badlyDistributedBlocks = 35;</code>
* @return The badlyDistributedBlocks.
*/
long getBadlyDistributedBlocks();
}
/**
* Protobuf type {@code hadoop.hdfs.NamenodeMembershipStatsRecordProto}
*/
public static final class NamenodeMembershipStatsRecordProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.NamenodeMembershipStatsRecordProto)
NamenodeMembershipStatsRecordProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use NamenodeMembershipStatsRecordProto.newBuilder() to construct.
private NamenodeMembershipStatsRecordProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private NamenodeMembershipStatsRecordProto() {
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new NamenodeMembershipStatsRecordProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_NamenodeMembershipStatsRecordProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_NamenodeMembershipStatsRecordProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto.Builder.class);
}
private int bitField0_;
public static final int TOTALSPACE_FIELD_NUMBER = 1;
private long totalSpace_ = 0L;
/**
* <code>optional uint64 totalSpace = 1;</code>
* @return Whether the totalSpace field is set.
*/
@java.lang.Override
public boolean hasTotalSpace() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional uint64 totalSpace = 1;</code>
* @return The totalSpace.
*/
@java.lang.Override
public long getTotalSpace() {
return totalSpace_;
}
public static final int AVAILABLESPACE_FIELD_NUMBER = 2;
private long availableSpace_ = 0L;
/**
* <code>optional uint64 availableSpace = 2;</code>
* @return Whether the availableSpace field is set.
*/
@java.lang.Override
public boolean hasAvailableSpace() {
return ((bitField0_ & 0x00000002) != 0);
}
/**
* <code>optional uint64 availableSpace = 2;</code>
* @return The availableSpace.
*/
@java.lang.Override
public long getAvailableSpace() {
return availableSpace_;
}
public static final int PROVIDEDSPACE_FIELD_NUMBER = 3;
private long providedSpace_ = 0L;
/**
* <code>optional uint64 providedSpace = 3;</code>
* @return Whether the providedSpace field is set.
*/
@java.lang.Override
public boolean hasProvidedSpace() {
return ((bitField0_ & 0x00000004) != 0);
}
/**
* <code>optional uint64 providedSpace = 3;</code>
* @return The providedSpace.
*/
@java.lang.Override
public long getProvidedSpace() {
return providedSpace_;
}
public static final int NUMOFFILES_FIELD_NUMBER = 10;
private long numOfFiles_ = 0L;
/**
* <code>optional uint64 numOfFiles = 10;</code>
* @return Whether the numOfFiles field is set.
*/
@java.lang.Override
public boolean hasNumOfFiles() {
return ((bitField0_ & 0x00000008) != 0);
}
/**
* <code>optional uint64 numOfFiles = 10;</code>
* @return The numOfFiles.
*/
@java.lang.Override
public long getNumOfFiles() {
return numOfFiles_;
}
public static final int NUMOFBLOCKS_FIELD_NUMBER = 11;
private long numOfBlocks_ = 0L;
/**
* <code>optional uint64 numOfBlocks = 11;</code>
* @return Whether the numOfBlocks field is set.
*/
@java.lang.Override
public boolean hasNumOfBlocks() {
return ((bitField0_ & 0x00000010) != 0);
}
/**
* <code>optional uint64 numOfBlocks = 11;</code>
* @return The numOfBlocks.
*/
@java.lang.Override
public long getNumOfBlocks() {
return numOfBlocks_;
}
public static final int NUMOFBLOCKSMISSING_FIELD_NUMBER = 12;
private long numOfBlocksMissing_ = 0L;
/**
* <code>optional uint64 numOfBlocksMissing = 12;</code>
* @return Whether the numOfBlocksMissing field is set.
*/
@java.lang.Override
public boolean hasNumOfBlocksMissing() {
return ((bitField0_ & 0x00000020) != 0);
}
/**
* <code>optional uint64 numOfBlocksMissing = 12;</code>
* @return The numOfBlocksMissing.
*/
@java.lang.Override
public long getNumOfBlocksMissing() {
return numOfBlocksMissing_;
}
public static final int NUMOFBLOCKSPENDINGREPLICATION_FIELD_NUMBER = 13;
private long numOfBlocksPendingReplication_ = 0L;
/**
* <code>optional uint64 numOfBlocksPendingReplication = 13;</code>
* @return Whether the numOfBlocksPendingReplication field is set.
*/
@java.lang.Override
public boolean hasNumOfBlocksPendingReplication() {
return ((bitField0_ & 0x00000040) != 0);
}
/**
* <code>optional uint64 numOfBlocksPendingReplication = 13;</code>
* @return The numOfBlocksPendingReplication.
*/
@java.lang.Override
public long getNumOfBlocksPendingReplication() {
return numOfBlocksPendingReplication_;
}
public static final int NUMOFBLOCKSUNDERREPLICATED_FIELD_NUMBER = 14;
private long numOfBlocksUnderReplicated_ = 0L;
/**
* <code>optional uint64 numOfBlocksUnderReplicated = 14;</code>
* @return Whether the numOfBlocksUnderReplicated field is set.
*/
@java.lang.Override
public boolean hasNumOfBlocksUnderReplicated() {
return ((bitField0_ & 0x00000080) != 0);
}
/**
* <code>optional uint64 numOfBlocksUnderReplicated = 14;</code>
* @return The numOfBlocksUnderReplicated.
*/
@java.lang.Override
public long getNumOfBlocksUnderReplicated() {
return numOfBlocksUnderReplicated_;
}
public static final int NUMOFBLOCKSPENDINGDELETION_FIELD_NUMBER = 15;
private long numOfBlocksPendingDeletion_ = 0L;
/**
* <code>optional uint64 numOfBlocksPendingDeletion = 15;</code>
* @return Whether the numOfBlocksPendingDeletion field is set.
*/
@java.lang.Override
public boolean hasNumOfBlocksPendingDeletion() {
return ((bitField0_ & 0x00000100) != 0);
}
/**
* <code>optional uint64 numOfBlocksPendingDeletion = 15;</code>
* @return The numOfBlocksPendingDeletion.
*/
@java.lang.Override
public long getNumOfBlocksPendingDeletion() {
return numOfBlocksPendingDeletion_;
}
public static final int NUMOFACTIVEDATANODES_FIELD_NUMBER = 20;
private int numOfActiveDatanodes_ = 0;
/**
* <code>optional uint32 numOfActiveDatanodes = 20;</code>
* @return Whether the numOfActiveDatanodes field is set.
*/
@java.lang.Override
public boolean hasNumOfActiveDatanodes() {
return ((bitField0_ & 0x00000200) != 0);
}
/**
* <code>optional uint32 numOfActiveDatanodes = 20;</code>
* @return The numOfActiveDatanodes.
*/
@java.lang.Override
public int getNumOfActiveDatanodes() {
return numOfActiveDatanodes_;
}
public static final int NUMOFDEADDATANODES_FIELD_NUMBER = 21;
private int numOfDeadDatanodes_ = 0;
/**
* <code>optional uint32 numOfDeadDatanodes = 21;</code>
* @return Whether the numOfDeadDatanodes field is set.
*/
@java.lang.Override
public boolean hasNumOfDeadDatanodes() {
return ((bitField0_ & 0x00000400) != 0);
}
/**
* <code>optional uint32 numOfDeadDatanodes = 21;</code>
* @return The numOfDeadDatanodes.
*/
@java.lang.Override
public int getNumOfDeadDatanodes() {
return numOfDeadDatanodes_;
}
public static final int NUMOFDECOMMISSIONINGDATANODES_FIELD_NUMBER = 22;
private int numOfDecommissioningDatanodes_ = 0;
/**
* <code>optional uint32 numOfDecommissioningDatanodes = 22;</code>
* @return Whether the numOfDecommissioningDatanodes field is set.
*/
@java.lang.Override
public boolean hasNumOfDecommissioningDatanodes() {
return ((bitField0_ & 0x00000800) != 0);
}
/**
* <code>optional uint32 numOfDecommissioningDatanodes = 22;</code>
* @return The numOfDecommissioningDatanodes.
*/
@java.lang.Override
public int getNumOfDecommissioningDatanodes() {
return numOfDecommissioningDatanodes_;
}
public static final int NUMOFDECOMACTIVEDATANODES_FIELD_NUMBER = 23;
private int numOfDecomActiveDatanodes_ = 0;
/**
* <code>optional uint32 numOfDecomActiveDatanodes = 23;</code>
* @return Whether the numOfDecomActiveDatanodes field is set.
*/
@java.lang.Override
public boolean hasNumOfDecomActiveDatanodes() {
return ((bitField0_ & 0x00001000) != 0);
}
/**
* <code>optional uint32 numOfDecomActiveDatanodes = 23;</code>
* @return The numOfDecomActiveDatanodes.
*/
@java.lang.Override
public int getNumOfDecomActiveDatanodes() {
return numOfDecomActiveDatanodes_;
}
public static final int NUMOFDECOMDEADDATANODES_FIELD_NUMBER = 24;
private int numOfDecomDeadDatanodes_ = 0;
/**
* <code>optional uint32 numOfDecomDeadDatanodes = 24;</code>
* @return Whether the numOfDecomDeadDatanodes field is set.
*/
@java.lang.Override
public boolean hasNumOfDecomDeadDatanodes() {
return ((bitField0_ & 0x00002000) != 0);
}
/**
* <code>optional uint32 numOfDecomDeadDatanodes = 24;</code>
* @return The numOfDecomDeadDatanodes.
*/
@java.lang.Override
public int getNumOfDecomDeadDatanodes() {
return numOfDecomDeadDatanodes_;
}
public static final int NUMOFSTALEDATANODES_FIELD_NUMBER = 25;
private int numOfStaleDatanodes_ = 0;
/**
* <code>optional uint32 numOfStaleDatanodes = 25;</code>
* @return Whether the numOfStaleDatanodes field is set.
*/
@java.lang.Override
public boolean hasNumOfStaleDatanodes() {
return ((bitField0_ & 0x00004000) != 0);
}
/**
* <code>optional uint32 numOfStaleDatanodes = 25;</code>
* @return The numOfStaleDatanodes.
*/
@java.lang.Override
public int getNumOfStaleDatanodes() {
return numOfStaleDatanodes_;
}
public static final int NUMOFINMAINTENANCELIVEDATANODES_FIELD_NUMBER = 26;
private int numOfInMaintenanceLiveDataNodes_ = 0;
/**
* <code>optional uint32 numOfInMaintenanceLiveDataNodes = 26;</code>
* @return Whether the numOfInMaintenanceLiveDataNodes field is set.
*/
@java.lang.Override
public boolean hasNumOfInMaintenanceLiveDataNodes() {
return ((bitField0_ & 0x00008000) != 0);
}
/**
* <code>optional uint32 numOfInMaintenanceLiveDataNodes = 26;</code>
* @return The numOfInMaintenanceLiveDataNodes.
*/
@java.lang.Override
public int getNumOfInMaintenanceLiveDataNodes() {
return numOfInMaintenanceLiveDataNodes_;
}
public static final int NUMOFINMAINTENANCEDEADDATANODES_FIELD_NUMBER = 27;
private int numOfInMaintenanceDeadDataNodes_ = 0;
/**
* <code>optional uint32 numOfInMaintenanceDeadDataNodes = 27;</code>
* @return Whether the numOfInMaintenanceDeadDataNodes field is set.
*/
@java.lang.Override
public boolean hasNumOfInMaintenanceDeadDataNodes() {
return ((bitField0_ & 0x00010000) != 0);
}
/**
* <code>optional uint32 numOfInMaintenanceDeadDataNodes = 27;</code>
* @return The numOfInMaintenanceDeadDataNodes.
*/
@java.lang.Override
public int getNumOfInMaintenanceDeadDataNodes() {
return numOfInMaintenanceDeadDataNodes_;
}
public static final int NUMOFENTERINGMAINTENANCEDATANODES_FIELD_NUMBER = 28;
private int numOfEnteringMaintenanceDataNodes_ = 0;
/**
* <code>optional uint32 numOfEnteringMaintenanceDataNodes = 28;</code>
* @return Whether the numOfEnteringMaintenanceDataNodes field is set.
*/
@java.lang.Override
public boolean hasNumOfEnteringMaintenanceDataNodes() {
return ((bitField0_ & 0x00020000) != 0);
}
/**
* <code>optional uint32 numOfEnteringMaintenanceDataNodes = 28;</code>
* @return The numOfEnteringMaintenanceDataNodes.
*/
@java.lang.Override
public int getNumOfEnteringMaintenanceDataNodes() {
return numOfEnteringMaintenanceDataNodes_;
}
public static final int CORRUPTFILESCOUNT_FIELD_NUMBER = 29;
private int corruptFilesCount_ = 0;
/**
* <code>optional uint32 corruptFilesCount = 29;</code>
* @return Whether the corruptFilesCount field is set.
*/
@java.lang.Override
public boolean hasCorruptFilesCount() {
return ((bitField0_ & 0x00040000) != 0);
}
/**
* <code>optional uint32 corruptFilesCount = 29;</code>
* @return The corruptFilesCount.
*/
@java.lang.Override
public int getCorruptFilesCount() {
return corruptFilesCount_;
}
public static final int SCHEDULEDREPLICATIONBLOCKS_FIELD_NUMBER = 30;
private long scheduledReplicationBlocks_ = 0L;
/**
* <code>optional uint64 scheduledReplicationBlocks = 30;</code>
* @return Whether the scheduledReplicationBlocks field is set.
*/
@java.lang.Override
public boolean hasScheduledReplicationBlocks() {
return ((bitField0_ & 0x00080000) != 0);
}
/**
* <code>optional uint64 scheduledReplicationBlocks = 30;</code>
* @return The scheduledReplicationBlocks.
*/
@java.lang.Override
public long getScheduledReplicationBlocks() {
return scheduledReplicationBlocks_;
}
public static final int NUMBEROFMISSINGBLOCKSWITHREPLICATIONFACTORONE_FIELD_NUMBER = 31;
private long numberOfMissingBlocksWithReplicationFactorOne_ = 0L;
/**
* <code>optional uint64 numberOfMissingBlocksWithReplicationFactorOne = 31;</code>
* @return Whether the numberOfMissingBlocksWithReplicationFactorOne field is set.
*/
@java.lang.Override
public boolean hasNumberOfMissingBlocksWithReplicationFactorOne() {
return ((bitField0_ & 0x00100000) != 0);
}
/**
* <code>optional uint64 numberOfMissingBlocksWithReplicationFactorOne = 31;</code>
* @return The numberOfMissingBlocksWithReplicationFactorOne.
*/
@java.lang.Override
public long getNumberOfMissingBlocksWithReplicationFactorOne() {
return numberOfMissingBlocksWithReplicationFactorOne_;
}
public static final int HIGHESTPRIORITYLOWREDUNDANCYREPLICATEDBLOCKS_FIELD_NUMBER = 32;
private long highestPriorityLowRedundancyReplicatedBlocks_ = 0L;
/**
* <code>optional uint64 highestPriorityLowRedundancyReplicatedBlocks = 32;</code>
* @return Whether the highestPriorityLowRedundancyReplicatedBlocks field is set.
*/
@java.lang.Override
public boolean hasHighestPriorityLowRedundancyReplicatedBlocks() {
return ((bitField0_ & 0x00200000) != 0);
}
/**
* <code>optional uint64 highestPriorityLowRedundancyReplicatedBlocks = 32;</code>
* @return The highestPriorityLowRedundancyReplicatedBlocks.
*/
@java.lang.Override
public long getHighestPriorityLowRedundancyReplicatedBlocks() {
return highestPriorityLowRedundancyReplicatedBlocks_;
}
public static final int HIGHESTPRIORITYLOWREDUNDANCYECBLOCKS_FIELD_NUMBER = 33;
private long highestPriorityLowRedundancyECBlocks_ = 0L;
/**
* <code>optional uint64 HighestPriorityLowRedundancyECBlocks = 33;</code>
* @return Whether the highestPriorityLowRedundancyECBlocks field is set.
*/
@java.lang.Override
public boolean hasHighestPriorityLowRedundancyECBlocks() {
return ((bitField0_ & 0x00400000) != 0);
}
/**
* <code>optional uint64 HighestPriorityLowRedundancyECBlocks = 33;</code>
* @return The highestPriorityLowRedundancyECBlocks.
*/
@java.lang.Override
public long getHighestPriorityLowRedundancyECBlocks() {
return highestPriorityLowRedundancyECBlocks_;
}
public static final int PENDINGSPSPATHS_FIELD_NUMBER = 34;
private int pendingSPSPaths_ = 0;
/**
* <code>optional uint32 pendingSPSPaths = 34;</code>
* @return Whether the pendingSPSPaths field is set.
*/
@java.lang.Override
public boolean hasPendingSPSPaths() {
return ((bitField0_ & 0x00800000) != 0);
}
/**
* <code>optional uint32 pendingSPSPaths = 34;</code>
* @return The pendingSPSPaths.
*/
@java.lang.Override
public int getPendingSPSPaths() {
return pendingSPSPaths_;
}
public static final int BADLYDISTRIBUTEDBLOCKS_FIELD_NUMBER = 35;
private long badlyDistributedBlocks_ = 0L;
/**
* <code>optional uint64 badlyDistributedBlocks = 35;</code>
* @return Whether the badlyDistributedBlocks field is set.
*/
@java.lang.Override
public boolean hasBadlyDistributedBlocks() {
return ((bitField0_ & 0x01000000) != 0);
}
/**
* <code>optional uint64 badlyDistributedBlocks = 35;</code>
* @return The badlyDistributedBlocks.
*/
@java.lang.Override
public long getBadlyDistributedBlocks() {
return badlyDistributedBlocks_;
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) != 0)) {
output.writeUInt64(1, totalSpace_);
}
if (((bitField0_ & 0x00000002) != 0)) {
output.writeUInt64(2, availableSpace_);
}
if (((bitField0_ & 0x00000004) != 0)) {
output.writeUInt64(3, providedSpace_);
}
if (((bitField0_ & 0x00000008) != 0)) {
output.writeUInt64(10, numOfFiles_);
}
if (((bitField0_ & 0x00000010) != 0)) {
output.writeUInt64(11, numOfBlocks_);
}
if (((bitField0_ & 0x00000020) != 0)) {
output.writeUInt64(12, numOfBlocksMissing_);
}
if (((bitField0_ & 0x00000040) != 0)) {
output.writeUInt64(13, numOfBlocksPendingReplication_);
}
if (((bitField0_ & 0x00000080) != 0)) {
output.writeUInt64(14, numOfBlocksUnderReplicated_);
}
if (((bitField0_ & 0x00000100) != 0)) {
output.writeUInt64(15, numOfBlocksPendingDeletion_);
}
if (((bitField0_ & 0x00000200) != 0)) {
output.writeUInt32(20, numOfActiveDatanodes_);
}
if (((bitField0_ & 0x00000400) != 0)) {
output.writeUInt32(21, numOfDeadDatanodes_);
}
if (((bitField0_ & 0x00000800) != 0)) {
output.writeUInt32(22, numOfDecommissioningDatanodes_);
}
if (((bitField0_ & 0x00001000) != 0)) {
output.writeUInt32(23, numOfDecomActiveDatanodes_);
}
if (((bitField0_ & 0x00002000) != 0)) {
output.writeUInt32(24, numOfDecomDeadDatanodes_);
}
if (((bitField0_ & 0x00004000) != 0)) {
output.writeUInt32(25, numOfStaleDatanodes_);
}
if (((bitField0_ & 0x00008000) != 0)) {
output.writeUInt32(26, numOfInMaintenanceLiveDataNodes_);
}
if (((bitField0_ & 0x00010000) != 0)) {
output.writeUInt32(27, numOfInMaintenanceDeadDataNodes_);
}
if (((bitField0_ & 0x00020000) != 0)) {
output.writeUInt32(28, numOfEnteringMaintenanceDataNodes_);
}
if (((bitField0_ & 0x00040000) != 0)) {
output.writeUInt32(29, corruptFilesCount_);
}
if (((bitField0_ & 0x00080000) != 0)) {
output.writeUInt64(30, scheduledReplicationBlocks_);
}
if (((bitField0_ & 0x00100000) != 0)) {
output.writeUInt64(31, numberOfMissingBlocksWithReplicationFactorOne_);
}
if (((bitField0_ & 0x00200000) != 0)) {
output.writeUInt64(32, highestPriorityLowRedundancyReplicatedBlocks_);
}
if (((bitField0_ & 0x00400000) != 0)) {
output.writeUInt64(33, highestPriorityLowRedundancyECBlocks_);
}
if (((bitField0_ & 0x00800000) != 0)) {
output.writeUInt32(34, pendingSPSPaths_);
}
if (((bitField0_ & 0x01000000) != 0)) {
output.writeUInt64(35, badlyDistributedBlocks_);
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeUInt64Size(1, totalSpace_);
}
if (((bitField0_ & 0x00000002) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeUInt64Size(2, availableSpace_);
}
if (((bitField0_ & 0x00000004) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeUInt64Size(3, providedSpace_);
}
if (((bitField0_ & 0x00000008) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeUInt64Size(10, numOfFiles_);
}
if (((bitField0_ & 0x00000010) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeUInt64Size(11, numOfBlocks_);
}
if (((bitField0_ & 0x00000020) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeUInt64Size(12, numOfBlocksMissing_);
}
if (((bitField0_ & 0x00000040) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeUInt64Size(13, numOfBlocksPendingReplication_);
}
if (((bitField0_ & 0x00000080) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeUInt64Size(14, numOfBlocksUnderReplicated_);
}
if (((bitField0_ & 0x00000100) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeUInt64Size(15, numOfBlocksPendingDeletion_);
}
if (((bitField0_ & 0x00000200) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeUInt32Size(20, numOfActiveDatanodes_);
}
if (((bitField0_ & 0x00000400) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeUInt32Size(21, numOfDeadDatanodes_);
}
if (((bitField0_ & 0x00000800) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeUInt32Size(22, numOfDecommissioningDatanodes_);
}
if (((bitField0_ & 0x00001000) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeUInt32Size(23, numOfDecomActiveDatanodes_);
}
if (((bitField0_ & 0x00002000) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeUInt32Size(24, numOfDecomDeadDatanodes_);
}
if (((bitField0_ & 0x00004000) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeUInt32Size(25, numOfStaleDatanodes_);
}
if (((bitField0_ & 0x00008000) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeUInt32Size(26, numOfInMaintenanceLiveDataNodes_);
}
if (((bitField0_ & 0x00010000) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeUInt32Size(27, numOfInMaintenanceDeadDataNodes_);
}
if (((bitField0_ & 0x00020000) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeUInt32Size(28, numOfEnteringMaintenanceDataNodes_);
}
if (((bitField0_ & 0x00040000) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeUInt32Size(29, corruptFilesCount_);
}
if (((bitField0_ & 0x00080000) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeUInt64Size(30, scheduledReplicationBlocks_);
}
if (((bitField0_ & 0x00100000) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeUInt64Size(31, numberOfMissingBlocksWithReplicationFactorOne_);
}
if (((bitField0_ & 0x00200000) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeUInt64Size(32, highestPriorityLowRedundancyReplicatedBlocks_);
}
if (((bitField0_ & 0x00400000) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeUInt64Size(33, highestPriorityLowRedundancyECBlocks_);
}
if (((bitField0_ & 0x00800000) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeUInt32Size(34, pendingSPSPaths_);
}
if (((bitField0_ & 0x01000000) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeUInt64Size(35, badlyDistributedBlocks_);
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto) obj;
if (hasTotalSpace() != other.hasTotalSpace()) return false;
if (hasTotalSpace()) {
if (getTotalSpace()
!= other.getTotalSpace()) return false;
}
if (hasAvailableSpace() != other.hasAvailableSpace()) return false;
if (hasAvailableSpace()) {
if (getAvailableSpace()
!= other.getAvailableSpace()) return false;
}
if (hasProvidedSpace() != other.hasProvidedSpace()) return false;
if (hasProvidedSpace()) {
if (getProvidedSpace()
!= other.getProvidedSpace()) return false;
}
if (hasNumOfFiles() != other.hasNumOfFiles()) return false;
if (hasNumOfFiles()) {
if (getNumOfFiles()
!= other.getNumOfFiles()) return false;
}
if (hasNumOfBlocks() != other.hasNumOfBlocks()) return false;
if (hasNumOfBlocks()) {
if (getNumOfBlocks()
!= other.getNumOfBlocks()) return false;
}
if (hasNumOfBlocksMissing() != other.hasNumOfBlocksMissing()) return false;
if (hasNumOfBlocksMissing()) {
if (getNumOfBlocksMissing()
!= other.getNumOfBlocksMissing()) return false;
}
if (hasNumOfBlocksPendingReplication() != other.hasNumOfBlocksPendingReplication()) return false;
if (hasNumOfBlocksPendingReplication()) {
if (getNumOfBlocksPendingReplication()
!= other.getNumOfBlocksPendingReplication()) return false;
}
if (hasNumOfBlocksUnderReplicated() != other.hasNumOfBlocksUnderReplicated()) return false;
if (hasNumOfBlocksUnderReplicated()) {
if (getNumOfBlocksUnderReplicated()
!= other.getNumOfBlocksUnderReplicated()) return false;
}
if (hasNumOfBlocksPendingDeletion() != other.hasNumOfBlocksPendingDeletion()) return false;
if (hasNumOfBlocksPendingDeletion()) {
if (getNumOfBlocksPendingDeletion()
!= other.getNumOfBlocksPendingDeletion()) return false;
}
if (hasNumOfActiveDatanodes() != other.hasNumOfActiveDatanodes()) return false;
if (hasNumOfActiveDatanodes()) {
if (getNumOfActiveDatanodes()
!= other.getNumOfActiveDatanodes()) return false;
}
if (hasNumOfDeadDatanodes() != other.hasNumOfDeadDatanodes()) return false;
if (hasNumOfDeadDatanodes()) {
if (getNumOfDeadDatanodes()
!= other.getNumOfDeadDatanodes()) return false;
}
if (hasNumOfDecommissioningDatanodes() != other.hasNumOfDecommissioningDatanodes()) return false;
if (hasNumOfDecommissioningDatanodes()) {
if (getNumOfDecommissioningDatanodes()
!= other.getNumOfDecommissioningDatanodes()) return false;
}
if (hasNumOfDecomActiveDatanodes() != other.hasNumOfDecomActiveDatanodes()) return false;
if (hasNumOfDecomActiveDatanodes()) {
if (getNumOfDecomActiveDatanodes()
!= other.getNumOfDecomActiveDatanodes()) return false;
}
if (hasNumOfDecomDeadDatanodes() != other.hasNumOfDecomDeadDatanodes()) return false;
if (hasNumOfDecomDeadDatanodes()) {
if (getNumOfDecomDeadDatanodes()
!= other.getNumOfDecomDeadDatanodes()) return false;
}
if (hasNumOfStaleDatanodes() != other.hasNumOfStaleDatanodes()) return false;
if (hasNumOfStaleDatanodes()) {
if (getNumOfStaleDatanodes()
!= other.getNumOfStaleDatanodes()) return false;
}
if (hasNumOfInMaintenanceLiveDataNodes() != other.hasNumOfInMaintenanceLiveDataNodes()) return false;
if (hasNumOfInMaintenanceLiveDataNodes()) {
if (getNumOfInMaintenanceLiveDataNodes()
!= other.getNumOfInMaintenanceLiveDataNodes()) return false;
}
if (hasNumOfInMaintenanceDeadDataNodes() != other.hasNumOfInMaintenanceDeadDataNodes()) return false;
if (hasNumOfInMaintenanceDeadDataNodes()) {
if (getNumOfInMaintenanceDeadDataNodes()
!= other.getNumOfInMaintenanceDeadDataNodes()) return false;
}
if (hasNumOfEnteringMaintenanceDataNodes() != other.hasNumOfEnteringMaintenanceDataNodes()) return false;
if (hasNumOfEnteringMaintenanceDataNodes()) {
if (getNumOfEnteringMaintenanceDataNodes()
!= other.getNumOfEnteringMaintenanceDataNodes()) return false;
}
if (hasCorruptFilesCount() != other.hasCorruptFilesCount()) return false;
if (hasCorruptFilesCount()) {
if (getCorruptFilesCount()
!= other.getCorruptFilesCount()) return false;
}
if (hasScheduledReplicationBlocks() != other.hasScheduledReplicationBlocks()) return false;
if (hasScheduledReplicationBlocks()) {
if (getScheduledReplicationBlocks()
!= other.getScheduledReplicationBlocks()) return false;
}
if (hasNumberOfMissingBlocksWithReplicationFactorOne() != other.hasNumberOfMissingBlocksWithReplicationFactorOne()) return false;
if (hasNumberOfMissingBlocksWithReplicationFactorOne()) {
if (getNumberOfMissingBlocksWithReplicationFactorOne()
!= other.getNumberOfMissingBlocksWithReplicationFactorOne()) return false;
}
if (hasHighestPriorityLowRedundancyReplicatedBlocks() != other.hasHighestPriorityLowRedundancyReplicatedBlocks()) return false;
if (hasHighestPriorityLowRedundancyReplicatedBlocks()) {
if (getHighestPriorityLowRedundancyReplicatedBlocks()
!= other.getHighestPriorityLowRedundancyReplicatedBlocks()) return false;
}
if (hasHighestPriorityLowRedundancyECBlocks() != other.hasHighestPriorityLowRedundancyECBlocks()) return false;
if (hasHighestPriorityLowRedundancyECBlocks()) {
if (getHighestPriorityLowRedundancyECBlocks()
!= other.getHighestPriorityLowRedundancyECBlocks()) return false;
}
if (hasPendingSPSPaths() != other.hasPendingSPSPaths()) return false;
if (hasPendingSPSPaths()) {
if (getPendingSPSPaths()
!= other.getPendingSPSPaths()) return false;
}
if (hasBadlyDistributedBlocks() != other.hasBadlyDistributedBlocks()) return false;
if (hasBadlyDistributedBlocks()) {
if (getBadlyDistributedBlocks()
!= other.getBadlyDistributedBlocks()) return false;
}
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (hasTotalSpace()) {
hash = (37 * hash) + TOTALSPACE_FIELD_NUMBER;
hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashLong(
getTotalSpace());
}
if (hasAvailableSpace()) {
hash = (37 * hash) + AVAILABLESPACE_FIELD_NUMBER;
hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashLong(
getAvailableSpace());
}
if (hasProvidedSpace()) {
hash = (37 * hash) + PROVIDEDSPACE_FIELD_NUMBER;
hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashLong(
getProvidedSpace());
}
if (hasNumOfFiles()) {
hash = (37 * hash) + NUMOFFILES_FIELD_NUMBER;
hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashLong(
getNumOfFiles());
}
if (hasNumOfBlocks()) {
hash = (37 * hash) + NUMOFBLOCKS_FIELD_NUMBER;
hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashLong(
getNumOfBlocks());
}
if (hasNumOfBlocksMissing()) {
hash = (37 * hash) + NUMOFBLOCKSMISSING_FIELD_NUMBER;
hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashLong(
getNumOfBlocksMissing());
}
if (hasNumOfBlocksPendingReplication()) {
hash = (37 * hash) + NUMOFBLOCKSPENDINGREPLICATION_FIELD_NUMBER;
hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashLong(
getNumOfBlocksPendingReplication());
}
if (hasNumOfBlocksUnderReplicated()) {
hash = (37 * hash) + NUMOFBLOCKSUNDERREPLICATED_FIELD_NUMBER;
hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashLong(
getNumOfBlocksUnderReplicated());
}
if (hasNumOfBlocksPendingDeletion()) {
hash = (37 * hash) + NUMOFBLOCKSPENDINGDELETION_FIELD_NUMBER;
hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashLong(
getNumOfBlocksPendingDeletion());
}
if (hasNumOfActiveDatanodes()) {
hash = (37 * hash) + NUMOFACTIVEDATANODES_FIELD_NUMBER;
hash = (53 * hash) + getNumOfActiveDatanodes();
}
if (hasNumOfDeadDatanodes()) {
hash = (37 * hash) + NUMOFDEADDATANODES_FIELD_NUMBER;
hash = (53 * hash) + getNumOfDeadDatanodes();
}
if (hasNumOfDecommissioningDatanodes()) {
hash = (37 * hash) + NUMOFDECOMMISSIONINGDATANODES_FIELD_NUMBER;
hash = (53 * hash) + getNumOfDecommissioningDatanodes();
}
if (hasNumOfDecomActiveDatanodes()) {
hash = (37 * hash) + NUMOFDECOMACTIVEDATANODES_FIELD_NUMBER;
hash = (53 * hash) + getNumOfDecomActiveDatanodes();
}
if (hasNumOfDecomDeadDatanodes()) {
hash = (37 * hash) + NUMOFDECOMDEADDATANODES_FIELD_NUMBER;
hash = (53 * hash) + getNumOfDecomDeadDatanodes();
}
if (hasNumOfStaleDatanodes()) {
hash = (37 * hash) + NUMOFSTALEDATANODES_FIELD_NUMBER;
hash = (53 * hash) + getNumOfStaleDatanodes();
}
if (hasNumOfInMaintenanceLiveDataNodes()) {
hash = (37 * hash) + NUMOFINMAINTENANCELIVEDATANODES_FIELD_NUMBER;
hash = (53 * hash) + getNumOfInMaintenanceLiveDataNodes();
}
if (hasNumOfInMaintenanceDeadDataNodes()) {
hash = (37 * hash) + NUMOFINMAINTENANCEDEADDATANODES_FIELD_NUMBER;
hash = (53 * hash) + getNumOfInMaintenanceDeadDataNodes();
}
if (hasNumOfEnteringMaintenanceDataNodes()) {
hash = (37 * hash) + NUMOFENTERINGMAINTENANCEDATANODES_FIELD_NUMBER;
hash = (53 * hash) + getNumOfEnteringMaintenanceDataNodes();
}
if (hasCorruptFilesCount()) {
hash = (37 * hash) + CORRUPTFILESCOUNT_FIELD_NUMBER;
hash = (53 * hash) + getCorruptFilesCount();
}
if (hasScheduledReplicationBlocks()) {
hash = (37 * hash) + SCHEDULEDREPLICATIONBLOCKS_FIELD_NUMBER;
hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashLong(
getScheduledReplicationBlocks());
}
if (hasNumberOfMissingBlocksWithReplicationFactorOne()) {
hash = (37 * hash) + NUMBEROFMISSINGBLOCKSWITHREPLICATIONFACTORONE_FIELD_NUMBER;
hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashLong(
getNumberOfMissingBlocksWithReplicationFactorOne());
}
if (hasHighestPriorityLowRedundancyReplicatedBlocks()) {
hash = (37 * hash) + HIGHESTPRIORITYLOWREDUNDANCYREPLICATEDBLOCKS_FIELD_NUMBER;
hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashLong(
getHighestPriorityLowRedundancyReplicatedBlocks());
}
if (hasHighestPriorityLowRedundancyECBlocks()) {
hash = (37 * hash) + HIGHESTPRIORITYLOWREDUNDANCYECBLOCKS_FIELD_NUMBER;
hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashLong(
getHighestPriorityLowRedundancyECBlocks());
}
if (hasPendingSPSPaths()) {
hash = (37 * hash) + PENDINGSPSPATHS_FIELD_NUMBER;
hash = (53 * hash) + getPendingSPSPaths();
}
if (hasBadlyDistributedBlocks()) {
hash = (37 * hash) + BADLYDISTRIBUTEDBLOCKS_FIELD_NUMBER;
hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashLong(
getBadlyDistributedBlocks());
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.NamenodeMembershipStatsRecordProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.NamenodeMembershipStatsRecordProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_NamenodeMembershipStatsRecordProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_NamenodeMembershipStatsRecordProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
totalSpace_ = 0L;
availableSpace_ = 0L;
providedSpace_ = 0L;
numOfFiles_ = 0L;
numOfBlocks_ = 0L;
numOfBlocksMissing_ = 0L;
numOfBlocksPendingReplication_ = 0L;
numOfBlocksUnderReplicated_ = 0L;
numOfBlocksPendingDeletion_ = 0L;
numOfActiveDatanodes_ = 0;
numOfDeadDatanodes_ = 0;
numOfDecommissioningDatanodes_ = 0;
numOfDecomActiveDatanodes_ = 0;
numOfDecomDeadDatanodes_ = 0;
numOfStaleDatanodes_ = 0;
numOfInMaintenanceLiveDataNodes_ = 0;
numOfInMaintenanceDeadDataNodes_ = 0;
numOfEnteringMaintenanceDataNodes_ = 0;
corruptFilesCount_ = 0;
scheduledReplicationBlocks_ = 0L;
numberOfMissingBlocksWithReplicationFactorOne_ = 0L;
highestPriorityLowRedundancyReplicatedBlocks_ = 0L;
highestPriorityLowRedundancyECBlocks_ = 0L;
pendingSPSPaths_ = 0;
badlyDistributedBlocks_ = 0L;
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_NamenodeMembershipStatsRecordProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto(this);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartial0(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto result) {
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) != 0)) {
result.totalSpace_ = totalSpace_;
to_bitField0_ |= 0x00000001;
}
if (((from_bitField0_ & 0x00000002) != 0)) {
result.availableSpace_ = availableSpace_;
to_bitField0_ |= 0x00000002;
}
if (((from_bitField0_ & 0x00000004) != 0)) {
result.providedSpace_ = providedSpace_;
to_bitField0_ |= 0x00000004;
}
if (((from_bitField0_ & 0x00000008) != 0)) {
result.numOfFiles_ = numOfFiles_;
to_bitField0_ |= 0x00000008;
}
if (((from_bitField0_ & 0x00000010) != 0)) {
result.numOfBlocks_ = numOfBlocks_;
to_bitField0_ |= 0x00000010;
}
if (((from_bitField0_ & 0x00000020) != 0)) {
result.numOfBlocksMissing_ = numOfBlocksMissing_;
to_bitField0_ |= 0x00000020;
}
if (((from_bitField0_ & 0x00000040) != 0)) {
result.numOfBlocksPendingReplication_ = numOfBlocksPendingReplication_;
to_bitField0_ |= 0x00000040;
}
if (((from_bitField0_ & 0x00000080) != 0)) {
result.numOfBlocksUnderReplicated_ = numOfBlocksUnderReplicated_;
to_bitField0_ |= 0x00000080;
}
if (((from_bitField0_ & 0x00000100) != 0)) {
result.numOfBlocksPendingDeletion_ = numOfBlocksPendingDeletion_;
to_bitField0_ |= 0x00000100;
}
if (((from_bitField0_ & 0x00000200) != 0)) {
result.numOfActiveDatanodes_ = numOfActiveDatanodes_;
to_bitField0_ |= 0x00000200;
}
if (((from_bitField0_ & 0x00000400) != 0)) {
result.numOfDeadDatanodes_ = numOfDeadDatanodes_;
to_bitField0_ |= 0x00000400;
}
if (((from_bitField0_ & 0x00000800) != 0)) {
result.numOfDecommissioningDatanodes_ = numOfDecommissioningDatanodes_;
to_bitField0_ |= 0x00000800;
}
if (((from_bitField0_ & 0x00001000) != 0)) {
result.numOfDecomActiveDatanodes_ = numOfDecomActiveDatanodes_;
to_bitField0_ |= 0x00001000;
}
if (((from_bitField0_ & 0x00002000) != 0)) {
result.numOfDecomDeadDatanodes_ = numOfDecomDeadDatanodes_;
to_bitField0_ |= 0x00002000;
}
if (((from_bitField0_ & 0x00004000) != 0)) {
result.numOfStaleDatanodes_ = numOfStaleDatanodes_;
to_bitField0_ |= 0x00004000;
}
if (((from_bitField0_ & 0x00008000) != 0)) {
result.numOfInMaintenanceLiveDataNodes_ = numOfInMaintenanceLiveDataNodes_;
to_bitField0_ |= 0x00008000;
}
if (((from_bitField0_ & 0x00010000) != 0)) {
result.numOfInMaintenanceDeadDataNodes_ = numOfInMaintenanceDeadDataNodes_;
to_bitField0_ |= 0x00010000;
}
if (((from_bitField0_ & 0x00020000) != 0)) {
result.numOfEnteringMaintenanceDataNodes_ = numOfEnteringMaintenanceDataNodes_;
to_bitField0_ |= 0x00020000;
}
if (((from_bitField0_ & 0x00040000) != 0)) {
result.corruptFilesCount_ = corruptFilesCount_;
to_bitField0_ |= 0x00040000;
}
if (((from_bitField0_ & 0x00080000) != 0)) {
result.scheduledReplicationBlocks_ = scheduledReplicationBlocks_;
to_bitField0_ |= 0x00080000;
}
if (((from_bitField0_ & 0x00100000) != 0)) {
result.numberOfMissingBlocksWithReplicationFactorOne_ = numberOfMissingBlocksWithReplicationFactorOne_;
to_bitField0_ |= 0x00100000;
}
if (((from_bitField0_ & 0x00200000) != 0)) {
result.highestPriorityLowRedundancyReplicatedBlocks_ = highestPriorityLowRedundancyReplicatedBlocks_;
to_bitField0_ |= 0x00200000;
}
if (((from_bitField0_ & 0x00400000) != 0)) {
result.highestPriorityLowRedundancyECBlocks_ = highestPriorityLowRedundancyECBlocks_;
to_bitField0_ |= 0x00400000;
}
if (((from_bitField0_ & 0x00800000) != 0)) {
result.pendingSPSPaths_ = pendingSPSPaths_;
to_bitField0_ |= 0x00800000;
}
if (((from_bitField0_ & 0x01000000) != 0)) {
result.badlyDistributedBlocks_ = badlyDistributedBlocks_;
to_bitField0_ |= 0x01000000;
}
result.bitField0_ |= to_bitField0_;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto.getDefaultInstance()) return this;
if (other.hasTotalSpace()) {
setTotalSpace(other.getTotalSpace());
}
if (other.hasAvailableSpace()) {
setAvailableSpace(other.getAvailableSpace());
}
if (other.hasProvidedSpace()) {
setProvidedSpace(other.getProvidedSpace());
}
if (other.hasNumOfFiles()) {
setNumOfFiles(other.getNumOfFiles());
}
if (other.hasNumOfBlocks()) {
setNumOfBlocks(other.getNumOfBlocks());
}
if (other.hasNumOfBlocksMissing()) {
setNumOfBlocksMissing(other.getNumOfBlocksMissing());
}
if (other.hasNumOfBlocksPendingReplication()) {
setNumOfBlocksPendingReplication(other.getNumOfBlocksPendingReplication());
}
if (other.hasNumOfBlocksUnderReplicated()) {
setNumOfBlocksUnderReplicated(other.getNumOfBlocksUnderReplicated());
}
if (other.hasNumOfBlocksPendingDeletion()) {
setNumOfBlocksPendingDeletion(other.getNumOfBlocksPendingDeletion());
}
if (other.hasNumOfActiveDatanodes()) {
setNumOfActiveDatanodes(other.getNumOfActiveDatanodes());
}
if (other.hasNumOfDeadDatanodes()) {
setNumOfDeadDatanodes(other.getNumOfDeadDatanodes());
}
if (other.hasNumOfDecommissioningDatanodes()) {
setNumOfDecommissioningDatanodes(other.getNumOfDecommissioningDatanodes());
}
if (other.hasNumOfDecomActiveDatanodes()) {
setNumOfDecomActiveDatanodes(other.getNumOfDecomActiveDatanodes());
}
if (other.hasNumOfDecomDeadDatanodes()) {
setNumOfDecomDeadDatanodes(other.getNumOfDecomDeadDatanodes());
}
if (other.hasNumOfStaleDatanodes()) {
setNumOfStaleDatanodes(other.getNumOfStaleDatanodes());
}
if (other.hasNumOfInMaintenanceLiveDataNodes()) {
setNumOfInMaintenanceLiveDataNodes(other.getNumOfInMaintenanceLiveDataNodes());
}
if (other.hasNumOfInMaintenanceDeadDataNodes()) {
setNumOfInMaintenanceDeadDataNodes(other.getNumOfInMaintenanceDeadDataNodes());
}
if (other.hasNumOfEnteringMaintenanceDataNodes()) {
setNumOfEnteringMaintenanceDataNodes(other.getNumOfEnteringMaintenanceDataNodes());
}
if (other.hasCorruptFilesCount()) {
setCorruptFilesCount(other.getCorruptFilesCount());
}
if (other.hasScheduledReplicationBlocks()) {
setScheduledReplicationBlocks(other.getScheduledReplicationBlocks());
}
if (other.hasNumberOfMissingBlocksWithReplicationFactorOne()) {
setNumberOfMissingBlocksWithReplicationFactorOne(other.getNumberOfMissingBlocksWithReplicationFactorOne());
}
if (other.hasHighestPriorityLowRedundancyReplicatedBlocks()) {
setHighestPriorityLowRedundancyReplicatedBlocks(other.getHighestPriorityLowRedundancyReplicatedBlocks());
}
if (other.hasHighestPriorityLowRedundancyECBlocks()) {
setHighestPriorityLowRedundancyECBlocks(other.getHighestPriorityLowRedundancyECBlocks());
}
if (other.hasPendingSPSPaths()) {
setPendingSPSPaths(other.getPendingSPSPaths());
}
if (other.hasBadlyDistributedBlocks()) {
setBadlyDistributedBlocks(other.getBadlyDistributedBlocks());
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 8: {
totalSpace_ = input.readUInt64();
bitField0_ |= 0x00000001;
break;
} // case 8
case 16: {
availableSpace_ = input.readUInt64();
bitField0_ |= 0x00000002;
break;
} // case 16
case 24: {
providedSpace_ = input.readUInt64();
bitField0_ |= 0x00000004;
break;
} // case 24
case 80: {
numOfFiles_ = input.readUInt64();
bitField0_ |= 0x00000008;
break;
} // case 80
case 88: {
numOfBlocks_ = input.readUInt64();
bitField0_ |= 0x00000010;
break;
} // case 88
case 96: {
numOfBlocksMissing_ = input.readUInt64();
bitField0_ |= 0x00000020;
break;
} // case 96
case 104: {
numOfBlocksPendingReplication_ = input.readUInt64();
bitField0_ |= 0x00000040;
break;
} // case 104
case 112: {
numOfBlocksUnderReplicated_ = input.readUInt64();
bitField0_ |= 0x00000080;
break;
} // case 112
case 120: {
numOfBlocksPendingDeletion_ = input.readUInt64();
bitField0_ |= 0x00000100;
break;
} // case 120
case 160: {
numOfActiveDatanodes_ = input.readUInt32();
bitField0_ |= 0x00000200;
break;
} // case 160
case 168: {
numOfDeadDatanodes_ = input.readUInt32();
bitField0_ |= 0x00000400;
break;
} // case 168
case 176: {
numOfDecommissioningDatanodes_ = input.readUInt32();
bitField0_ |= 0x00000800;
break;
} // case 176
case 184: {
numOfDecomActiveDatanodes_ = input.readUInt32();
bitField0_ |= 0x00001000;
break;
} // case 184
case 192: {
numOfDecomDeadDatanodes_ = input.readUInt32();
bitField0_ |= 0x00002000;
break;
} // case 192
case 200: {
numOfStaleDatanodes_ = input.readUInt32();
bitField0_ |= 0x00004000;
break;
} // case 200
case 208: {
numOfInMaintenanceLiveDataNodes_ = input.readUInt32();
bitField0_ |= 0x00008000;
break;
} // case 208
case 216: {
numOfInMaintenanceDeadDataNodes_ = input.readUInt32();
bitField0_ |= 0x00010000;
break;
} // case 216
case 224: {
numOfEnteringMaintenanceDataNodes_ = input.readUInt32();
bitField0_ |= 0x00020000;
break;
} // case 224
case 232: {
corruptFilesCount_ = input.readUInt32();
bitField0_ |= 0x00040000;
break;
} // case 232
case 240: {
scheduledReplicationBlocks_ = input.readUInt64();
bitField0_ |= 0x00080000;
break;
} // case 240
case 248: {
numberOfMissingBlocksWithReplicationFactorOne_ = input.readUInt64();
bitField0_ |= 0x00100000;
break;
} // case 248
case 256: {
highestPriorityLowRedundancyReplicatedBlocks_ = input.readUInt64();
bitField0_ |= 0x00200000;
break;
} // case 256
case 264: {
highestPriorityLowRedundancyECBlocks_ = input.readUInt64();
bitField0_ |= 0x00400000;
break;
} // case 264
case 272: {
pendingSPSPaths_ = input.readUInt32();
bitField0_ |= 0x00800000;
break;
} // case 272
case 280: {
badlyDistributedBlocks_ = input.readUInt64();
bitField0_ |= 0x01000000;
break;
} // case 280
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private long totalSpace_ ;
/**
* <code>optional uint64 totalSpace = 1;</code>
* @return Whether the totalSpace field is set.
*/
@java.lang.Override
public boolean hasTotalSpace() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional uint64 totalSpace = 1;</code>
* @return The totalSpace.
*/
@java.lang.Override
public long getTotalSpace() {
return totalSpace_;
}
/**
* <code>optional uint64 totalSpace = 1;</code>
* @param value The totalSpace to set.
* @return This builder for chaining.
*/
public Builder setTotalSpace(long value) {
totalSpace_ = value;
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>optional uint64 totalSpace = 1;</code>
* @return This builder for chaining.
*/
public Builder clearTotalSpace() {
bitField0_ = (bitField0_ & ~0x00000001);
totalSpace_ = 0L;
onChanged();
return this;
}
private long availableSpace_ ;
/**
* <code>optional uint64 availableSpace = 2;</code>
* @return Whether the availableSpace field is set.
*/
@java.lang.Override
public boolean hasAvailableSpace() {
return ((bitField0_ & 0x00000002) != 0);
}
/**
* <code>optional uint64 availableSpace = 2;</code>
* @return The availableSpace.
*/
@java.lang.Override
public long getAvailableSpace() {
return availableSpace_;
}
/**
* <code>optional uint64 availableSpace = 2;</code>
* @param value The availableSpace to set.
* @return This builder for chaining.
*/
public Builder setAvailableSpace(long value) {
availableSpace_ = value;
bitField0_ |= 0x00000002;
onChanged();
return this;
}
/**
* <code>optional uint64 availableSpace = 2;</code>
* @return This builder for chaining.
*/
public Builder clearAvailableSpace() {
bitField0_ = (bitField0_ & ~0x00000002);
availableSpace_ = 0L;
onChanged();
return this;
}
private long providedSpace_ ;
/**
* <code>optional uint64 providedSpace = 3;</code>
* @return Whether the providedSpace field is set.
*/
@java.lang.Override
public boolean hasProvidedSpace() {
return ((bitField0_ & 0x00000004) != 0);
}
/**
* <code>optional uint64 providedSpace = 3;</code>
* @return The providedSpace.
*/
@java.lang.Override
public long getProvidedSpace() {
return providedSpace_;
}
/**
* <code>optional uint64 providedSpace = 3;</code>
* @param value The providedSpace to set.
* @return This builder for chaining.
*/
public Builder setProvidedSpace(long value) {
providedSpace_ = value;
bitField0_ |= 0x00000004;
onChanged();
return this;
}
/**
* <code>optional uint64 providedSpace = 3;</code>
* @return This builder for chaining.
*/
public Builder clearProvidedSpace() {
bitField0_ = (bitField0_ & ~0x00000004);
providedSpace_ = 0L;
onChanged();
return this;
}
private long numOfFiles_ ;
/**
* <code>optional uint64 numOfFiles = 10;</code>
* @return Whether the numOfFiles field is set.
*/
@java.lang.Override
public boolean hasNumOfFiles() {
return ((bitField0_ & 0x00000008) != 0);
}
/**
* <code>optional uint64 numOfFiles = 10;</code>
* @return The numOfFiles.
*/
@java.lang.Override
public long getNumOfFiles() {
return numOfFiles_;
}
/**
* <code>optional uint64 numOfFiles = 10;</code>
* @param value The numOfFiles to set.
* @return This builder for chaining.
*/
public Builder setNumOfFiles(long value) {
numOfFiles_ = value;
bitField0_ |= 0x00000008;
onChanged();
return this;
}
/**
* <code>optional uint64 numOfFiles = 10;</code>
* @return This builder for chaining.
*/
public Builder clearNumOfFiles() {
bitField0_ = (bitField0_ & ~0x00000008);
numOfFiles_ = 0L;
onChanged();
return this;
}
private long numOfBlocks_ ;
/**
* <code>optional uint64 numOfBlocks = 11;</code>
* @return Whether the numOfBlocks field is set.
*/
@java.lang.Override
public boolean hasNumOfBlocks() {
return ((bitField0_ & 0x00000010) != 0);
}
/**
* <code>optional uint64 numOfBlocks = 11;</code>
* @return The numOfBlocks.
*/
@java.lang.Override
public long getNumOfBlocks() {
return numOfBlocks_;
}
/**
* <code>optional uint64 numOfBlocks = 11;</code>
* @param value The numOfBlocks to set.
* @return This builder for chaining.
*/
public Builder setNumOfBlocks(long value) {
numOfBlocks_ = value;
bitField0_ |= 0x00000010;
onChanged();
return this;
}
/**
* <code>optional uint64 numOfBlocks = 11;</code>
* @return This builder for chaining.
*/
public Builder clearNumOfBlocks() {
bitField0_ = (bitField0_ & ~0x00000010);
numOfBlocks_ = 0L;
onChanged();
return this;
}
private long numOfBlocksMissing_ ;
/**
* <code>optional uint64 numOfBlocksMissing = 12;</code>
* @return Whether the numOfBlocksMissing field is set.
*/
@java.lang.Override
public boolean hasNumOfBlocksMissing() {
return ((bitField0_ & 0x00000020) != 0);
}
/**
* <code>optional uint64 numOfBlocksMissing = 12;</code>
* @return The numOfBlocksMissing.
*/
@java.lang.Override
public long getNumOfBlocksMissing() {
return numOfBlocksMissing_;
}
/**
* <code>optional uint64 numOfBlocksMissing = 12;</code>
* @param value The numOfBlocksMissing to set.
* @return This builder for chaining.
*/
public Builder setNumOfBlocksMissing(long value) {
numOfBlocksMissing_ = value;
bitField0_ |= 0x00000020;
onChanged();
return this;
}
/**
* <code>optional uint64 numOfBlocksMissing = 12;</code>
* @return This builder for chaining.
*/
public Builder clearNumOfBlocksMissing() {
bitField0_ = (bitField0_ & ~0x00000020);
numOfBlocksMissing_ = 0L;
onChanged();
return this;
}
private long numOfBlocksPendingReplication_ ;
/**
* <code>optional uint64 numOfBlocksPendingReplication = 13;</code>
* @return Whether the numOfBlocksPendingReplication field is set.
*/
@java.lang.Override
public boolean hasNumOfBlocksPendingReplication() {
return ((bitField0_ & 0x00000040) != 0);
}
/**
* <code>optional uint64 numOfBlocksPendingReplication = 13;</code>
* @return The numOfBlocksPendingReplication.
*/
@java.lang.Override
public long getNumOfBlocksPendingReplication() {
return numOfBlocksPendingReplication_;
}
/**
* <code>optional uint64 numOfBlocksPendingReplication = 13;</code>
* @param value The numOfBlocksPendingReplication to set.
* @return This builder for chaining.
*/
public Builder setNumOfBlocksPendingReplication(long value) {
numOfBlocksPendingReplication_ = value;
bitField0_ |= 0x00000040;
onChanged();
return this;
}
/**
* <code>optional uint64 numOfBlocksPendingReplication = 13;</code>
* @return This builder for chaining.
*/
public Builder clearNumOfBlocksPendingReplication() {
bitField0_ = (bitField0_ & ~0x00000040);
numOfBlocksPendingReplication_ = 0L;
onChanged();
return this;
}
private long numOfBlocksUnderReplicated_ ;
/**
* <code>optional uint64 numOfBlocksUnderReplicated = 14;</code>
* @return Whether the numOfBlocksUnderReplicated field is set.
*/
@java.lang.Override
public boolean hasNumOfBlocksUnderReplicated() {
return ((bitField0_ & 0x00000080) != 0);
}
/**
* <code>optional uint64 numOfBlocksUnderReplicated = 14;</code>
* @return The numOfBlocksUnderReplicated.
*/
@java.lang.Override
public long getNumOfBlocksUnderReplicated() {
return numOfBlocksUnderReplicated_;
}
/**
* <code>optional uint64 numOfBlocksUnderReplicated = 14;</code>
* @param value The numOfBlocksUnderReplicated to set.
* @return This builder for chaining.
*/
public Builder setNumOfBlocksUnderReplicated(long value) {
numOfBlocksUnderReplicated_ = value;
bitField0_ |= 0x00000080;
onChanged();
return this;
}
/**
* <code>optional uint64 numOfBlocksUnderReplicated = 14;</code>
* @return This builder for chaining.
*/
public Builder clearNumOfBlocksUnderReplicated() {
bitField0_ = (bitField0_ & ~0x00000080);
numOfBlocksUnderReplicated_ = 0L;
onChanged();
return this;
}
private long numOfBlocksPendingDeletion_ ;
/**
* <code>optional uint64 numOfBlocksPendingDeletion = 15;</code>
* @return Whether the numOfBlocksPendingDeletion field is set.
*/
@java.lang.Override
public boolean hasNumOfBlocksPendingDeletion() {
return ((bitField0_ & 0x00000100) != 0);
}
/**
* <code>optional uint64 numOfBlocksPendingDeletion = 15;</code>
* @return The numOfBlocksPendingDeletion.
*/
@java.lang.Override
public long getNumOfBlocksPendingDeletion() {
return numOfBlocksPendingDeletion_;
}
/**
* <code>optional uint64 numOfBlocksPendingDeletion = 15;</code>
* @param value The numOfBlocksPendingDeletion to set.
* @return This builder for chaining.
*/
public Builder setNumOfBlocksPendingDeletion(long value) {
numOfBlocksPendingDeletion_ = value;
bitField0_ |= 0x00000100;
onChanged();
return this;
}
/**
* <code>optional uint64 numOfBlocksPendingDeletion = 15;</code>
* @return This builder for chaining.
*/
public Builder clearNumOfBlocksPendingDeletion() {
bitField0_ = (bitField0_ & ~0x00000100);
numOfBlocksPendingDeletion_ = 0L;
onChanged();
return this;
}
private int numOfActiveDatanodes_ ;
/**
* <code>optional uint32 numOfActiveDatanodes = 20;</code>
* @return Whether the numOfActiveDatanodes field is set.
*/
@java.lang.Override
public boolean hasNumOfActiveDatanodes() {
return ((bitField0_ & 0x00000200) != 0);
}
/**
* <code>optional uint32 numOfActiveDatanodes = 20;</code>
* @return The numOfActiveDatanodes.
*/
@java.lang.Override
public int getNumOfActiveDatanodes() {
return numOfActiveDatanodes_;
}
/**
* <code>optional uint32 numOfActiveDatanodes = 20;</code>
* @param value The numOfActiveDatanodes to set.
* @return This builder for chaining.
*/
public Builder setNumOfActiveDatanodes(int value) {
numOfActiveDatanodes_ = value;
bitField0_ |= 0x00000200;
onChanged();
return this;
}
/**
* <code>optional uint32 numOfActiveDatanodes = 20;</code>
* @return This builder for chaining.
*/
public Builder clearNumOfActiveDatanodes() {
bitField0_ = (bitField0_ & ~0x00000200);
numOfActiveDatanodes_ = 0;
onChanged();
return this;
}
private int numOfDeadDatanodes_ ;
/**
* <code>optional uint32 numOfDeadDatanodes = 21;</code>
* @return Whether the numOfDeadDatanodes field is set.
*/
@java.lang.Override
public boolean hasNumOfDeadDatanodes() {
return ((bitField0_ & 0x00000400) != 0);
}
/**
* <code>optional uint32 numOfDeadDatanodes = 21;</code>
* @return The numOfDeadDatanodes.
*/
@java.lang.Override
public int getNumOfDeadDatanodes() {
return numOfDeadDatanodes_;
}
/**
* <code>optional uint32 numOfDeadDatanodes = 21;</code>
* @param value The numOfDeadDatanodes to set.
* @return This builder for chaining.
*/
public Builder setNumOfDeadDatanodes(int value) {
numOfDeadDatanodes_ = value;
bitField0_ |= 0x00000400;
onChanged();
return this;
}
/**
* <code>optional uint32 numOfDeadDatanodes = 21;</code>
* @return This builder for chaining.
*/
public Builder clearNumOfDeadDatanodes() {
bitField0_ = (bitField0_ & ~0x00000400);
numOfDeadDatanodes_ = 0;
onChanged();
return this;
}
private int numOfDecommissioningDatanodes_ ;
/**
* <code>optional uint32 numOfDecommissioningDatanodes = 22;</code>
* @return Whether the numOfDecommissioningDatanodes field is set.
*/
@java.lang.Override
public boolean hasNumOfDecommissioningDatanodes() {
return ((bitField0_ & 0x00000800) != 0);
}
/**
* <code>optional uint32 numOfDecommissioningDatanodes = 22;</code>
* @return The numOfDecommissioningDatanodes.
*/
@java.lang.Override
public int getNumOfDecommissioningDatanodes() {
return numOfDecommissioningDatanodes_;
}
/**
* <code>optional uint32 numOfDecommissioningDatanodes = 22;</code>
* @param value The numOfDecommissioningDatanodes to set.
* @return This builder for chaining.
*/
public Builder setNumOfDecommissioningDatanodes(int value) {
numOfDecommissioningDatanodes_ = value;
bitField0_ |= 0x00000800;
onChanged();
return this;
}
/**
* <code>optional uint32 numOfDecommissioningDatanodes = 22;</code>
* @return This builder for chaining.
*/
public Builder clearNumOfDecommissioningDatanodes() {
bitField0_ = (bitField0_ & ~0x00000800);
numOfDecommissioningDatanodes_ = 0;
onChanged();
return this;
}
private int numOfDecomActiveDatanodes_ ;
/**
* <code>optional uint32 numOfDecomActiveDatanodes = 23;</code>
* @return Whether the numOfDecomActiveDatanodes field is set.
*/
@java.lang.Override
public boolean hasNumOfDecomActiveDatanodes() {
return ((bitField0_ & 0x00001000) != 0);
}
/**
* <code>optional uint32 numOfDecomActiveDatanodes = 23;</code>
* @return The numOfDecomActiveDatanodes.
*/
@java.lang.Override
public int getNumOfDecomActiveDatanodes() {
return numOfDecomActiveDatanodes_;
}
/**
* <code>optional uint32 numOfDecomActiveDatanodes = 23;</code>
* @param value The numOfDecomActiveDatanodes to set.
* @return This builder for chaining.
*/
public Builder setNumOfDecomActiveDatanodes(int value) {
numOfDecomActiveDatanodes_ = value;
bitField0_ |= 0x00001000;
onChanged();
return this;
}
/**
* <code>optional uint32 numOfDecomActiveDatanodes = 23;</code>
* @return This builder for chaining.
*/
public Builder clearNumOfDecomActiveDatanodes() {
bitField0_ = (bitField0_ & ~0x00001000);
numOfDecomActiveDatanodes_ = 0;
onChanged();
return this;
}
private int numOfDecomDeadDatanodes_ ;
/**
* <code>optional uint32 numOfDecomDeadDatanodes = 24;</code>
* @return Whether the numOfDecomDeadDatanodes field is set.
*/
@java.lang.Override
public boolean hasNumOfDecomDeadDatanodes() {
return ((bitField0_ & 0x00002000) != 0);
}
/**
* <code>optional uint32 numOfDecomDeadDatanodes = 24;</code>
* @return The numOfDecomDeadDatanodes.
*/
@java.lang.Override
public int getNumOfDecomDeadDatanodes() {
return numOfDecomDeadDatanodes_;
}
/**
* <code>optional uint32 numOfDecomDeadDatanodes = 24;</code>
* @param value The numOfDecomDeadDatanodes to set.
* @return This builder for chaining.
*/
public Builder setNumOfDecomDeadDatanodes(int value) {
numOfDecomDeadDatanodes_ = value;
bitField0_ |= 0x00002000;
onChanged();
return this;
}
/**
* <code>optional uint32 numOfDecomDeadDatanodes = 24;</code>
* @return This builder for chaining.
*/
public Builder clearNumOfDecomDeadDatanodes() {
bitField0_ = (bitField0_ & ~0x00002000);
numOfDecomDeadDatanodes_ = 0;
onChanged();
return this;
}
private int numOfStaleDatanodes_ ;
/**
* <code>optional uint32 numOfStaleDatanodes = 25;</code>
* @return Whether the numOfStaleDatanodes field is set.
*/
@java.lang.Override
public boolean hasNumOfStaleDatanodes() {
return ((bitField0_ & 0x00004000) != 0);
}
/**
* <code>optional uint32 numOfStaleDatanodes = 25;</code>
* @return The numOfStaleDatanodes.
*/
@java.lang.Override
public int getNumOfStaleDatanodes() {
return numOfStaleDatanodes_;
}
/**
* <code>optional uint32 numOfStaleDatanodes = 25;</code>
* @param value The numOfStaleDatanodes to set.
* @return This builder for chaining.
*/
public Builder setNumOfStaleDatanodes(int value) {
numOfStaleDatanodes_ = value;
bitField0_ |= 0x00004000;
onChanged();
return this;
}
/**
* <code>optional uint32 numOfStaleDatanodes = 25;</code>
* @return This builder for chaining.
*/
public Builder clearNumOfStaleDatanodes() {
bitField0_ = (bitField0_ & ~0x00004000);
numOfStaleDatanodes_ = 0;
onChanged();
return this;
}
private int numOfInMaintenanceLiveDataNodes_ ;
/**
* <code>optional uint32 numOfInMaintenanceLiveDataNodes = 26;</code>
* @return Whether the numOfInMaintenanceLiveDataNodes field is set.
*/
@java.lang.Override
public boolean hasNumOfInMaintenanceLiveDataNodes() {
return ((bitField0_ & 0x00008000) != 0);
}
/**
* <code>optional uint32 numOfInMaintenanceLiveDataNodes = 26;</code>
* @return The numOfInMaintenanceLiveDataNodes.
*/
@java.lang.Override
public int getNumOfInMaintenanceLiveDataNodes() {
return numOfInMaintenanceLiveDataNodes_;
}
/**
* <code>optional uint32 numOfInMaintenanceLiveDataNodes = 26;</code>
* @param value The numOfInMaintenanceLiveDataNodes to set.
* @return This builder for chaining.
*/
public Builder setNumOfInMaintenanceLiveDataNodes(int value) {
numOfInMaintenanceLiveDataNodes_ = value;
bitField0_ |= 0x00008000;
onChanged();
return this;
}
/**
* <code>optional uint32 numOfInMaintenanceLiveDataNodes = 26;</code>
* @return This builder for chaining.
*/
public Builder clearNumOfInMaintenanceLiveDataNodes() {
bitField0_ = (bitField0_ & ~0x00008000);
numOfInMaintenanceLiveDataNodes_ = 0;
onChanged();
return this;
}
private int numOfInMaintenanceDeadDataNodes_ ;
/**
* <code>optional uint32 numOfInMaintenanceDeadDataNodes = 27;</code>
* @return Whether the numOfInMaintenanceDeadDataNodes field is set.
*/
@java.lang.Override
public boolean hasNumOfInMaintenanceDeadDataNodes() {
return ((bitField0_ & 0x00010000) != 0);
}
/**
* <code>optional uint32 numOfInMaintenanceDeadDataNodes = 27;</code>
* @return The numOfInMaintenanceDeadDataNodes.
*/
@java.lang.Override
public int getNumOfInMaintenanceDeadDataNodes() {
return numOfInMaintenanceDeadDataNodes_;
}
/**
* <code>optional uint32 numOfInMaintenanceDeadDataNodes = 27;</code>
* @param value The numOfInMaintenanceDeadDataNodes to set.
* @return This builder for chaining.
*/
public Builder setNumOfInMaintenanceDeadDataNodes(int value) {
numOfInMaintenanceDeadDataNodes_ = value;
bitField0_ |= 0x00010000;
onChanged();
return this;
}
/**
* <code>optional uint32 numOfInMaintenanceDeadDataNodes = 27;</code>
* @return This builder for chaining.
*/
public Builder clearNumOfInMaintenanceDeadDataNodes() {
bitField0_ = (bitField0_ & ~0x00010000);
numOfInMaintenanceDeadDataNodes_ = 0;
onChanged();
return this;
}
private int numOfEnteringMaintenanceDataNodes_ ;
/**
* <code>optional uint32 numOfEnteringMaintenanceDataNodes = 28;</code>
* @return Whether the numOfEnteringMaintenanceDataNodes field is set.
*/
@java.lang.Override
public boolean hasNumOfEnteringMaintenanceDataNodes() {
return ((bitField0_ & 0x00020000) != 0);
}
/**
* <code>optional uint32 numOfEnteringMaintenanceDataNodes = 28;</code>
* @return The numOfEnteringMaintenanceDataNodes.
*/
@java.lang.Override
public int getNumOfEnteringMaintenanceDataNodes() {
return numOfEnteringMaintenanceDataNodes_;
}
/**
* <code>optional uint32 numOfEnteringMaintenanceDataNodes = 28;</code>
* @param value The numOfEnteringMaintenanceDataNodes to set.
* @return This builder for chaining.
*/
public Builder setNumOfEnteringMaintenanceDataNodes(int value) {
numOfEnteringMaintenanceDataNodes_ = value;
bitField0_ |= 0x00020000;
onChanged();
return this;
}
/**
* <code>optional uint32 numOfEnteringMaintenanceDataNodes = 28;</code>
* @return This builder for chaining.
*/
public Builder clearNumOfEnteringMaintenanceDataNodes() {
bitField0_ = (bitField0_ & ~0x00020000);
numOfEnteringMaintenanceDataNodes_ = 0;
onChanged();
return this;
}
private int corruptFilesCount_ ;
/**
* <code>optional uint32 corruptFilesCount = 29;</code>
* @return Whether the corruptFilesCount field is set.
*/
@java.lang.Override
public boolean hasCorruptFilesCount() {
return ((bitField0_ & 0x00040000) != 0);
}
/**
* <code>optional uint32 corruptFilesCount = 29;</code>
* @return The corruptFilesCount.
*/
@java.lang.Override
public int getCorruptFilesCount() {
return corruptFilesCount_;
}
/**
* <code>optional uint32 corruptFilesCount = 29;</code>
* @param value The corruptFilesCount to set.
* @return This builder for chaining.
*/
public Builder setCorruptFilesCount(int value) {
corruptFilesCount_ = value;
bitField0_ |= 0x00040000;
onChanged();
return this;
}
/**
* <code>optional uint32 corruptFilesCount = 29;</code>
* @return This builder for chaining.
*/
public Builder clearCorruptFilesCount() {
bitField0_ = (bitField0_ & ~0x00040000);
corruptFilesCount_ = 0;
onChanged();
return this;
}
private long scheduledReplicationBlocks_ ;
/**
* <code>optional uint64 scheduledReplicationBlocks = 30;</code>
* @return Whether the scheduledReplicationBlocks field is set.
*/
@java.lang.Override
public boolean hasScheduledReplicationBlocks() {
return ((bitField0_ & 0x00080000) != 0);
}
/**
* <code>optional uint64 scheduledReplicationBlocks = 30;</code>
* @return The scheduledReplicationBlocks.
*/
@java.lang.Override
public long getScheduledReplicationBlocks() {
return scheduledReplicationBlocks_;
}
/**
* <code>optional uint64 scheduledReplicationBlocks = 30;</code>
* @param value The scheduledReplicationBlocks to set.
* @return This builder for chaining.
*/
public Builder setScheduledReplicationBlocks(long value) {
scheduledReplicationBlocks_ = value;
bitField0_ |= 0x00080000;
onChanged();
return this;
}
/**
* <code>optional uint64 scheduledReplicationBlocks = 30;</code>
* @return This builder for chaining.
*/
public Builder clearScheduledReplicationBlocks() {
bitField0_ = (bitField0_ & ~0x00080000);
scheduledReplicationBlocks_ = 0L;
onChanged();
return this;
}
private long numberOfMissingBlocksWithReplicationFactorOne_ ;
/**
* <code>optional uint64 numberOfMissingBlocksWithReplicationFactorOne = 31;</code>
* @return Whether the numberOfMissingBlocksWithReplicationFactorOne field is set.
*/
@java.lang.Override
public boolean hasNumberOfMissingBlocksWithReplicationFactorOne() {
return ((bitField0_ & 0x00100000) != 0);
}
/**
* <code>optional uint64 numberOfMissingBlocksWithReplicationFactorOne = 31;</code>
* @return The numberOfMissingBlocksWithReplicationFactorOne.
*/
@java.lang.Override
public long getNumberOfMissingBlocksWithReplicationFactorOne() {
return numberOfMissingBlocksWithReplicationFactorOne_;
}
/**
* <code>optional uint64 numberOfMissingBlocksWithReplicationFactorOne = 31;</code>
* @param value The numberOfMissingBlocksWithReplicationFactorOne to set.
* @return This builder for chaining.
*/
public Builder setNumberOfMissingBlocksWithReplicationFactorOne(long value) {
numberOfMissingBlocksWithReplicationFactorOne_ = value;
bitField0_ |= 0x00100000;
onChanged();
return this;
}
/**
* <code>optional uint64 numberOfMissingBlocksWithReplicationFactorOne = 31;</code>
* @return This builder for chaining.
*/
public Builder clearNumberOfMissingBlocksWithReplicationFactorOne() {
bitField0_ = (bitField0_ & ~0x00100000);
numberOfMissingBlocksWithReplicationFactorOne_ = 0L;
onChanged();
return this;
}
private long highestPriorityLowRedundancyReplicatedBlocks_ ;
/**
* <code>optional uint64 highestPriorityLowRedundancyReplicatedBlocks = 32;</code>
* @return Whether the highestPriorityLowRedundancyReplicatedBlocks field is set.
*/
@java.lang.Override
public boolean hasHighestPriorityLowRedundancyReplicatedBlocks() {
return ((bitField0_ & 0x00200000) != 0);
}
/**
* <code>optional uint64 highestPriorityLowRedundancyReplicatedBlocks = 32;</code>
* @return The highestPriorityLowRedundancyReplicatedBlocks.
*/
@java.lang.Override
public long getHighestPriorityLowRedundancyReplicatedBlocks() {
return highestPriorityLowRedundancyReplicatedBlocks_;
}
/**
* <code>optional uint64 highestPriorityLowRedundancyReplicatedBlocks = 32;</code>
* @param value The highestPriorityLowRedundancyReplicatedBlocks to set.
* @return This builder for chaining.
*/
public Builder setHighestPriorityLowRedundancyReplicatedBlocks(long value) {
highestPriorityLowRedundancyReplicatedBlocks_ = value;
bitField0_ |= 0x00200000;
onChanged();
return this;
}
/**
* <code>optional uint64 highestPriorityLowRedundancyReplicatedBlocks = 32;</code>
* @return This builder for chaining.
*/
public Builder clearHighestPriorityLowRedundancyReplicatedBlocks() {
bitField0_ = (bitField0_ & ~0x00200000);
highestPriorityLowRedundancyReplicatedBlocks_ = 0L;
onChanged();
return this;
}
private long highestPriorityLowRedundancyECBlocks_ ;
/**
* <code>optional uint64 HighestPriorityLowRedundancyECBlocks = 33;</code>
* @return Whether the highestPriorityLowRedundancyECBlocks field is set.
*/
@java.lang.Override
public boolean hasHighestPriorityLowRedundancyECBlocks() {
return ((bitField0_ & 0x00400000) != 0);
}
/**
* <code>optional uint64 HighestPriorityLowRedundancyECBlocks = 33;</code>
* @return The highestPriorityLowRedundancyECBlocks.
*/
@java.lang.Override
public long getHighestPriorityLowRedundancyECBlocks() {
return highestPriorityLowRedundancyECBlocks_;
}
/**
* <code>optional uint64 HighestPriorityLowRedundancyECBlocks = 33;</code>
* @param value The highestPriorityLowRedundancyECBlocks to set.
* @return This builder for chaining.
*/
public Builder setHighestPriorityLowRedundancyECBlocks(long value) {
highestPriorityLowRedundancyECBlocks_ = value;
bitField0_ |= 0x00400000;
onChanged();
return this;
}
/**
* <code>optional uint64 HighestPriorityLowRedundancyECBlocks = 33;</code>
* @return This builder for chaining.
*/
public Builder clearHighestPriorityLowRedundancyECBlocks() {
bitField0_ = (bitField0_ & ~0x00400000);
highestPriorityLowRedundancyECBlocks_ = 0L;
onChanged();
return this;
}
private int pendingSPSPaths_ ;
/**
* <code>optional uint32 pendingSPSPaths = 34;</code>
* @return Whether the pendingSPSPaths field is set.
*/
@java.lang.Override
public boolean hasPendingSPSPaths() {
return ((bitField0_ & 0x00800000) != 0);
}
/**
* <code>optional uint32 pendingSPSPaths = 34;</code>
* @return The pendingSPSPaths.
*/
@java.lang.Override
public int getPendingSPSPaths() {
return pendingSPSPaths_;
}
/**
* <code>optional uint32 pendingSPSPaths = 34;</code>
* @param value The pendingSPSPaths to set.
* @return This builder for chaining.
*/
public Builder setPendingSPSPaths(int value) {
pendingSPSPaths_ = value;
bitField0_ |= 0x00800000;
onChanged();
return this;
}
/**
* <code>optional uint32 pendingSPSPaths = 34;</code>
* @return This builder for chaining.
*/
public Builder clearPendingSPSPaths() {
bitField0_ = (bitField0_ & ~0x00800000);
pendingSPSPaths_ = 0;
onChanged();
return this;
}
private long badlyDistributedBlocks_ ;
/**
* <code>optional uint64 badlyDistributedBlocks = 35;</code>
* @return Whether the badlyDistributedBlocks field is set.
*/
@java.lang.Override
public boolean hasBadlyDistributedBlocks() {
return ((bitField0_ & 0x01000000) != 0);
}
/**
* <code>optional uint64 badlyDistributedBlocks = 35;</code>
* @return The badlyDistributedBlocks.
*/
@java.lang.Override
public long getBadlyDistributedBlocks() {
return badlyDistributedBlocks_;
}
/**
* <code>optional uint64 badlyDistributedBlocks = 35;</code>
* @param value The badlyDistributedBlocks to set.
* @return This builder for chaining.
*/
public Builder setBadlyDistributedBlocks(long value) {
badlyDistributedBlocks_ = value;
bitField0_ |= 0x01000000;
onChanged();
return this;
}
/**
* <code>optional uint64 badlyDistributedBlocks = 35;</code>
* @return This builder for chaining.
*/
public Builder clearBadlyDistributedBlocks() {
bitField0_ = (bitField0_ & ~0x01000000);
badlyDistributedBlocks_ = 0L;
onChanged();
return this;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.NamenodeMembershipStatsRecordProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.NamenodeMembershipStatsRecordProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<NamenodeMembershipStatsRecordProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<NamenodeMembershipStatsRecordProto>() {
@java.lang.Override
public NamenodeMembershipStatsRecordProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<NamenodeMembershipStatsRecordProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<NamenodeMembershipStatsRecordProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface NamenodeMembershipRecordProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.NamenodeMembershipRecordProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* <code>optional uint64 dateCreated = 1;</code>
* @return Whether the dateCreated field is set.
*/
boolean hasDateCreated();
/**
* <code>optional uint64 dateCreated = 1;</code>
* @return The dateCreated.
*/
long getDateCreated();
/**
* <code>optional uint64 dateModified = 2;</code>
* @return Whether the dateModified field is set.
*/
boolean hasDateModified();
/**
* <code>optional uint64 dateModified = 2;</code>
* @return The dateModified.
*/
long getDateModified();
/**
* <code>optional uint64 lastContact = 3;</code>
* @return Whether the lastContact field is set.
*/
boolean hasLastContact();
/**
* <code>optional uint64 lastContact = 3;</code>
* @return The lastContact.
*/
long getLastContact();
/**
* <code>optional string routerId = 4;</code>
* @return Whether the routerId field is set.
*/
boolean hasRouterId();
/**
* <code>optional string routerId = 4;</code>
* @return The routerId.
*/
java.lang.String getRouterId();
/**
* <code>optional string routerId = 4;</code>
* @return The bytes for routerId.
*/
org.apache.hadoop.thirdparty.protobuf.ByteString
getRouterIdBytes();
/**
* <code>optional string nameserviceId = 5;</code>
* @return Whether the nameserviceId field is set.
*/
boolean hasNameserviceId();
/**
* <code>optional string nameserviceId = 5;</code>
* @return The nameserviceId.
*/
java.lang.String getNameserviceId();
/**
* <code>optional string nameserviceId = 5;</code>
* @return The bytes for nameserviceId.
*/
org.apache.hadoop.thirdparty.protobuf.ByteString
getNameserviceIdBytes();
/**
* <code>optional string namenodeId = 6;</code>
* @return Whether the namenodeId field is set.
*/
boolean hasNamenodeId();
/**
* <code>optional string namenodeId = 6;</code>
* @return The namenodeId.
*/
java.lang.String getNamenodeId();
/**
* <code>optional string namenodeId = 6;</code>
* @return The bytes for namenodeId.
*/
org.apache.hadoop.thirdparty.protobuf.ByteString
getNamenodeIdBytes();
/**
* <code>optional string clusterId = 7;</code>
* @return Whether the clusterId field is set.
*/
boolean hasClusterId();
/**
* <code>optional string clusterId = 7;</code>
* @return The clusterId.
*/
java.lang.String getClusterId();
/**
* <code>optional string clusterId = 7;</code>
* @return The bytes for clusterId.
*/
org.apache.hadoop.thirdparty.protobuf.ByteString
getClusterIdBytes();
/**
* <code>optional string blockPoolId = 8;</code>
* @return Whether the blockPoolId field is set.
*/
boolean hasBlockPoolId();
/**
* <code>optional string blockPoolId = 8;</code>
* @return The blockPoolId.
*/
java.lang.String getBlockPoolId();
/**
* <code>optional string blockPoolId = 8;</code>
* @return The bytes for blockPoolId.
*/
org.apache.hadoop.thirdparty.protobuf.ByteString
getBlockPoolIdBytes();
/**
* <code>optional string webAddress = 9;</code>
* @return Whether the webAddress field is set.
*/
boolean hasWebAddress();
/**
* <code>optional string webAddress = 9;</code>
* @return The webAddress.
*/
java.lang.String getWebAddress();
/**
* <code>optional string webAddress = 9;</code>
* @return The bytes for webAddress.
*/
org.apache.hadoop.thirdparty.protobuf.ByteString
getWebAddressBytes();
/**
* <code>optional string rpcAddress = 10;</code>
* @return Whether the rpcAddress field is set.
*/
boolean hasRpcAddress();
/**
* <code>optional string rpcAddress = 10;</code>
* @return The rpcAddress.
*/
java.lang.String getRpcAddress();
/**
* <code>optional string rpcAddress = 10;</code>
* @return The bytes for rpcAddress.
*/
org.apache.hadoop.thirdparty.protobuf.ByteString
getRpcAddressBytes();
/**
* <code>optional string serviceAddress = 11;</code>
* @return Whether the serviceAddress field is set.
*/
boolean hasServiceAddress();
/**
* <code>optional string serviceAddress = 11;</code>
* @return The serviceAddress.
*/
java.lang.String getServiceAddress();
/**
* <code>optional string serviceAddress = 11;</code>
* @return The bytes for serviceAddress.
*/
org.apache.hadoop.thirdparty.protobuf.ByteString
getServiceAddressBytes();
/**
* <code>optional string lifelineAddress = 12;</code>
* @return Whether the lifelineAddress field is set.
*/
boolean hasLifelineAddress();
/**
* <code>optional string lifelineAddress = 12;</code>
* @return The lifelineAddress.
*/
java.lang.String getLifelineAddress();
/**
* <code>optional string lifelineAddress = 12;</code>
* @return The bytes for lifelineAddress.
*/
org.apache.hadoop.thirdparty.protobuf.ByteString
getLifelineAddressBytes();
/**
* <code>optional string state = 13;</code>
* @return Whether the state field is set.
*/
boolean hasState();
/**
* <code>optional string state = 13;</code>
* @return The state.
*/
java.lang.String getState();
/**
* <code>optional string state = 13;</code>
* @return The bytes for state.
*/
org.apache.hadoop.thirdparty.protobuf.ByteString
getStateBytes();
/**
* <code>optional bool isSafeMode = 14;</code>
* @return Whether the isSafeMode field is set.
*/
boolean hasIsSafeMode();
/**
* <code>optional bool isSafeMode = 14;</code>
* @return The isSafeMode.
*/
boolean getIsSafeMode();
/**
* <code>optional .hadoop.hdfs.NamenodeMembershipStatsRecordProto stats = 15;</code>
* @return Whether the stats field is set.
*/
boolean hasStats();
/**
* <code>optional .hadoop.hdfs.NamenodeMembershipStatsRecordProto stats = 15;</code>
* @return The stats.
*/
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto getStats();
/**
* <code>optional .hadoop.hdfs.NamenodeMembershipStatsRecordProto stats = 15;</code>
*/
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProtoOrBuilder getStatsOrBuilder();
/**
* <code>optional string webScheme = 16;</code>
* @return Whether the webScheme field is set.
*/
boolean hasWebScheme();
/**
* <code>optional string webScheme = 16;</code>
* @return The webScheme.
*/
java.lang.String getWebScheme();
/**
* <code>optional string webScheme = 16;</code>
* @return The bytes for webScheme.
*/
org.apache.hadoop.thirdparty.protobuf.ByteString
getWebSchemeBytes();
}
/**
* Protobuf type {@code hadoop.hdfs.NamenodeMembershipRecordProto}
*/
public static final class NamenodeMembershipRecordProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.NamenodeMembershipRecordProto)
NamenodeMembershipRecordProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use NamenodeMembershipRecordProto.newBuilder() to construct.
private NamenodeMembershipRecordProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private NamenodeMembershipRecordProto() {
routerId_ = "";
nameserviceId_ = "";
namenodeId_ = "";
clusterId_ = "";
blockPoolId_ = "";
webAddress_ = "";
rpcAddress_ = "";
serviceAddress_ = "";
lifelineAddress_ = "";
state_ = "";
webScheme_ = "";
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new NamenodeMembershipRecordProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_NamenodeMembershipRecordProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_NamenodeMembershipRecordProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto.Builder.class);
}
private int bitField0_;
public static final int DATECREATED_FIELD_NUMBER = 1;
private long dateCreated_ = 0L;
/**
* <code>optional uint64 dateCreated = 1;</code>
* @return Whether the dateCreated field is set.
*/
@java.lang.Override
public boolean hasDateCreated() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional uint64 dateCreated = 1;</code>
* @return The dateCreated.
*/
@java.lang.Override
public long getDateCreated() {
return dateCreated_;
}
public static final int DATEMODIFIED_FIELD_NUMBER = 2;
private long dateModified_ = 0L;
/**
* <code>optional uint64 dateModified = 2;</code>
* @return Whether the dateModified field is set.
*/
@java.lang.Override
public boolean hasDateModified() {
return ((bitField0_ & 0x00000002) != 0);
}
/**
* <code>optional uint64 dateModified = 2;</code>
* @return The dateModified.
*/
@java.lang.Override
public long getDateModified() {
return dateModified_;
}
public static final int LASTCONTACT_FIELD_NUMBER = 3;
private long lastContact_ = 0L;
/**
* <code>optional uint64 lastContact = 3;</code>
* @return Whether the lastContact field is set.
*/
@java.lang.Override
public boolean hasLastContact() {
return ((bitField0_ & 0x00000004) != 0);
}
/**
* <code>optional uint64 lastContact = 3;</code>
* @return The lastContact.
*/
@java.lang.Override
public long getLastContact() {
return lastContact_;
}
public static final int ROUTERID_FIELD_NUMBER = 4;
@SuppressWarnings("serial")
private volatile java.lang.Object routerId_ = "";
/**
* <code>optional string routerId = 4;</code>
* @return Whether the routerId field is set.
*/
@java.lang.Override
public boolean hasRouterId() {
return ((bitField0_ & 0x00000008) != 0);
}
/**
* <code>optional string routerId = 4;</code>
* @return The routerId.
*/
@java.lang.Override
public java.lang.String getRouterId() {
java.lang.Object ref = routerId_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
routerId_ = s;
}
return s;
}
}
/**
* <code>optional string routerId = 4;</code>
* @return The bytes for routerId.
*/
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.ByteString
getRouterIdBytes() {
java.lang.Object ref = routerId_;
if (ref instanceof java.lang.String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
routerId_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
public static final int NAMESERVICEID_FIELD_NUMBER = 5;
@SuppressWarnings("serial")
private volatile java.lang.Object nameserviceId_ = "";
/**
* <code>optional string nameserviceId = 5;</code>
* @return Whether the nameserviceId field is set.
*/
@java.lang.Override
public boolean hasNameserviceId() {
return ((bitField0_ & 0x00000010) != 0);
}
/**
* <code>optional string nameserviceId = 5;</code>
* @return The nameserviceId.
*/
@java.lang.Override
public java.lang.String getNameserviceId() {
java.lang.Object ref = nameserviceId_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
nameserviceId_ = s;
}
return s;
}
}
/**
* <code>optional string nameserviceId = 5;</code>
* @return The bytes for nameserviceId.
*/
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.ByteString
getNameserviceIdBytes() {
java.lang.Object ref = nameserviceId_;
if (ref instanceof java.lang.String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
nameserviceId_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
public static final int NAMENODEID_FIELD_NUMBER = 6;
@SuppressWarnings("serial")
private volatile java.lang.Object namenodeId_ = "";
/**
* <code>optional string namenodeId = 6;</code>
* @return Whether the namenodeId field is set.
*/
@java.lang.Override
public boolean hasNamenodeId() {
return ((bitField0_ & 0x00000020) != 0);
}
/**
* <code>optional string namenodeId = 6;</code>
* @return The namenodeId.
*/
@java.lang.Override
public java.lang.String getNamenodeId() {
java.lang.Object ref = namenodeId_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
namenodeId_ = s;
}
return s;
}
}
/**
* <code>optional string namenodeId = 6;</code>
* @return The bytes for namenodeId.
*/
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.ByteString
getNamenodeIdBytes() {
java.lang.Object ref = namenodeId_;
if (ref instanceof java.lang.String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
namenodeId_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
public static final int CLUSTERID_FIELD_NUMBER = 7;
@SuppressWarnings("serial")
private volatile java.lang.Object clusterId_ = "";
/**
* <code>optional string clusterId = 7;</code>
* @return Whether the clusterId field is set.
*/
@java.lang.Override
public boolean hasClusterId() {
return ((bitField0_ & 0x00000040) != 0);
}
/**
* <code>optional string clusterId = 7;</code>
* @return The clusterId.
*/
@java.lang.Override
public java.lang.String getClusterId() {
java.lang.Object ref = clusterId_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
clusterId_ = s;
}
return s;
}
}
/**
* <code>optional string clusterId = 7;</code>
* @return The bytes for clusterId.
*/
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.ByteString
getClusterIdBytes() {
java.lang.Object ref = clusterId_;
if (ref instanceof java.lang.String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
clusterId_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
public static final int BLOCKPOOLID_FIELD_NUMBER = 8;
@SuppressWarnings("serial")
private volatile java.lang.Object blockPoolId_ = "";
/**
* <code>optional string blockPoolId = 8;</code>
* @return Whether the blockPoolId field is set.
*/
@java.lang.Override
public boolean hasBlockPoolId() {
return ((bitField0_ & 0x00000080) != 0);
}
/**
* <code>optional string blockPoolId = 8;</code>
* @return The blockPoolId.
*/
@java.lang.Override
public java.lang.String getBlockPoolId() {
java.lang.Object ref = blockPoolId_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
blockPoolId_ = s;
}
return s;
}
}
/**
* <code>optional string blockPoolId = 8;</code>
* @return The bytes for blockPoolId.
*/
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.ByteString
getBlockPoolIdBytes() {
java.lang.Object ref = blockPoolId_;
if (ref instanceof java.lang.String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
blockPoolId_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
public static final int WEBADDRESS_FIELD_NUMBER = 9;
@SuppressWarnings("serial")
private volatile java.lang.Object webAddress_ = "";
/**
* <code>optional string webAddress = 9;</code>
* @return Whether the webAddress field is set.
*/
@java.lang.Override
public boolean hasWebAddress() {
return ((bitField0_ & 0x00000100) != 0);
}
/**
* <code>optional string webAddress = 9;</code>
* @return The webAddress.
*/
@java.lang.Override
public java.lang.String getWebAddress() {
java.lang.Object ref = webAddress_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
webAddress_ = s;
}
return s;
}
}
/**
* <code>optional string webAddress = 9;</code>
* @return The bytes for webAddress.
*/
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.ByteString
getWebAddressBytes() {
java.lang.Object ref = webAddress_;
if (ref instanceof java.lang.String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
webAddress_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
public static final int RPCADDRESS_FIELD_NUMBER = 10;
@SuppressWarnings("serial")
private volatile java.lang.Object rpcAddress_ = "";
/**
* <code>optional string rpcAddress = 10;</code>
* @return Whether the rpcAddress field is set.
*/
@java.lang.Override
public boolean hasRpcAddress() {
return ((bitField0_ & 0x00000200) != 0);
}
/**
* <code>optional string rpcAddress = 10;</code>
* @return The rpcAddress.
*/
@java.lang.Override
public java.lang.String getRpcAddress() {
java.lang.Object ref = rpcAddress_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
rpcAddress_ = s;
}
return s;
}
}
/**
* <code>optional string rpcAddress = 10;</code>
* @return The bytes for rpcAddress.
*/
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.ByteString
getRpcAddressBytes() {
java.lang.Object ref = rpcAddress_;
if (ref instanceof java.lang.String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
rpcAddress_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
public static final int SERVICEADDRESS_FIELD_NUMBER = 11;
@SuppressWarnings("serial")
private volatile java.lang.Object serviceAddress_ = "";
/**
* <code>optional string serviceAddress = 11;</code>
* @return Whether the serviceAddress field is set.
*/
@java.lang.Override
public boolean hasServiceAddress() {
return ((bitField0_ & 0x00000400) != 0);
}
/**
* <code>optional string serviceAddress = 11;</code>
* @return The serviceAddress.
*/
@java.lang.Override
public java.lang.String getServiceAddress() {
java.lang.Object ref = serviceAddress_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
serviceAddress_ = s;
}
return s;
}
}
/**
* <code>optional string serviceAddress = 11;</code>
* @return The bytes for serviceAddress.
*/
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.ByteString
getServiceAddressBytes() {
java.lang.Object ref = serviceAddress_;
if (ref instanceof java.lang.String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
serviceAddress_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
public static final int LIFELINEADDRESS_FIELD_NUMBER = 12;
@SuppressWarnings("serial")
private volatile java.lang.Object lifelineAddress_ = "";
/**
* <code>optional string lifelineAddress = 12;</code>
* @return Whether the lifelineAddress field is set.
*/
@java.lang.Override
public boolean hasLifelineAddress() {
return ((bitField0_ & 0x00000800) != 0);
}
/**
* <code>optional string lifelineAddress = 12;</code>
* @return The lifelineAddress.
*/
@java.lang.Override
public java.lang.String getLifelineAddress() {
java.lang.Object ref = lifelineAddress_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
lifelineAddress_ = s;
}
return s;
}
}
/**
* <code>optional string lifelineAddress = 12;</code>
* @return The bytes for lifelineAddress.
*/
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.ByteString
getLifelineAddressBytes() {
java.lang.Object ref = lifelineAddress_;
if (ref instanceof java.lang.String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
lifelineAddress_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
public static final int STATE_FIELD_NUMBER = 13;
@SuppressWarnings("serial")
private volatile java.lang.Object state_ = "";
/**
* <code>optional string state = 13;</code>
* @return Whether the state field is set.
*/
@java.lang.Override
public boolean hasState() {
return ((bitField0_ & 0x00001000) != 0);
}
/**
* <code>optional string state = 13;</code>
* @return The state.
*/
@java.lang.Override
public java.lang.String getState() {
java.lang.Object ref = state_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
state_ = s;
}
return s;
}
}
/**
* <code>optional string state = 13;</code>
* @return The bytes for state.
*/
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.ByteString
getStateBytes() {
java.lang.Object ref = state_;
if (ref instanceof java.lang.String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
state_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
public static final int ISSAFEMODE_FIELD_NUMBER = 14;
private boolean isSafeMode_ = false;
/**
* <code>optional bool isSafeMode = 14;</code>
* @return Whether the isSafeMode field is set.
*/
@java.lang.Override
public boolean hasIsSafeMode() {
return ((bitField0_ & 0x00002000) != 0);
}
/**
* <code>optional bool isSafeMode = 14;</code>
* @return The isSafeMode.
*/
@java.lang.Override
public boolean getIsSafeMode() {
return isSafeMode_;
}
public static final int STATS_FIELD_NUMBER = 15;
private org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto stats_;
/**
* <code>optional .hadoop.hdfs.NamenodeMembershipStatsRecordProto stats = 15;</code>
* @return Whether the stats field is set.
*/
@java.lang.Override
public boolean hasStats() {
return ((bitField0_ & 0x00004000) != 0);
}
/**
* <code>optional .hadoop.hdfs.NamenodeMembershipStatsRecordProto stats = 15;</code>
* @return The stats.
*/
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto getStats() {
return stats_ == null ? org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto.getDefaultInstance() : stats_;
}
/**
* <code>optional .hadoop.hdfs.NamenodeMembershipStatsRecordProto stats = 15;</code>
*/
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProtoOrBuilder getStatsOrBuilder() {
return stats_ == null ? org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto.getDefaultInstance() : stats_;
}
public static final int WEBSCHEME_FIELD_NUMBER = 16;
@SuppressWarnings("serial")
private volatile java.lang.Object webScheme_ = "";
/**
* <code>optional string webScheme = 16;</code>
* @return Whether the webScheme field is set.
*/
@java.lang.Override
public boolean hasWebScheme() {
return ((bitField0_ & 0x00008000) != 0);
}
/**
* <code>optional string webScheme = 16;</code>
* @return The webScheme.
*/
@java.lang.Override
public java.lang.String getWebScheme() {
java.lang.Object ref = webScheme_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
webScheme_ = s;
}
return s;
}
}
/**
* <code>optional string webScheme = 16;</code>
* @return The bytes for webScheme.
*/
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.ByteString
getWebSchemeBytes() {
java.lang.Object ref = webScheme_;
if (ref instanceof java.lang.String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
webScheme_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) != 0)) {
output.writeUInt64(1, dateCreated_);
}
if (((bitField0_ & 0x00000002) != 0)) {
output.writeUInt64(2, dateModified_);
}
if (((bitField0_ & 0x00000004) != 0)) {
output.writeUInt64(3, lastContact_);
}
if (((bitField0_ & 0x00000008) != 0)) {
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 4, routerId_);
}
if (((bitField0_ & 0x00000010) != 0)) {
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 5, nameserviceId_);
}
if (((bitField0_ & 0x00000020) != 0)) {
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 6, namenodeId_);
}
if (((bitField0_ & 0x00000040) != 0)) {
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 7, clusterId_);
}
if (((bitField0_ & 0x00000080) != 0)) {
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 8, blockPoolId_);
}
if (((bitField0_ & 0x00000100) != 0)) {
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 9, webAddress_);
}
if (((bitField0_ & 0x00000200) != 0)) {
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 10, rpcAddress_);
}
if (((bitField0_ & 0x00000400) != 0)) {
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 11, serviceAddress_);
}
if (((bitField0_ & 0x00000800) != 0)) {
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 12, lifelineAddress_);
}
if (((bitField0_ & 0x00001000) != 0)) {
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 13, state_);
}
if (((bitField0_ & 0x00002000) != 0)) {
output.writeBool(14, isSafeMode_);
}
if (((bitField0_ & 0x00004000) != 0)) {
output.writeMessage(15, getStats());
}
if (((bitField0_ & 0x00008000) != 0)) {
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 16, webScheme_);
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeUInt64Size(1, dateCreated_);
}
if (((bitField0_ & 0x00000002) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeUInt64Size(2, dateModified_);
}
if (((bitField0_ & 0x00000004) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeUInt64Size(3, lastContact_);
}
if (((bitField0_ & 0x00000008) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(4, routerId_);
}
if (((bitField0_ & 0x00000010) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(5, nameserviceId_);
}
if (((bitField0_ & 0x00000020) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(6, namenodeId_);
}
if (((bitField0_ & 0x00000040) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(7, clusterId_);
}
if (((bitField0_ & 0x00000080) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(8, blockPoolId_);
}
if (((bitField0_ & 0x00000100) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(9, webAddress_);
}
if (((bitField0_ & 0x00000200) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(10, rpcAddress_);
}
if (((bitField0_ & 0x00000400) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(11, serviceAddress_);
}
if (((bitField0_ & 0x00000800) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(12, lifelineAddress_);
}
if (((bitField0_ & 0x00001000) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(13, state_);
}
if (((bitField0_ & 0x00002000) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeBoolSize(14, isSafeMode_);
}
if (((bitField0_ & 0x00004000) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeMessageSize(15, getStats());
}
if (((bitField0_ & 0x00008000) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(16, webScheme_);
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto) obj;
if (hasDateCreated() != other.hasDateCreated()) return false;
if (hasDateCreated()) {
if (getDateCreated()
!= other.getDateCreated()) return false;
}
if (hasDateModified() != other.hasDateModified()) return false;
if (hasDateModified()) {
if (getDateModified()
!= other.getDateModified()) return false;
}
if (hasLastContact() != other.hasLastContact()) return false;
if (hasLastContact()) {
if (getLastContact()
!= other.getLastContact()) return false;
}
if (hasRouterId() != other.hasRouterId()) return false;
if (hasRouterId()) {
if (!getRouterId()
.equals(other.getRouterId())) return false;
}
if (hasNameserviceId() != other.hasNameserviceId()) return false;
if (hasNameserviceId()) {
if (!getNameserviceId()
.equals(other.getNameserviceId())) return false;
}
if (hasNamenodeId() != other.hasNamenodeId()) return false;
if (hasNamenodeId()) {
if (!getNamenodeId()
.equals(other.getNamenodeId())) return false;
}
if (hasClusterId() != other.hasClusterId()) return false;
if (hasClusterId()) {
if (!getClusterId()
.equals(other.getClusterId())) return false;
}
if (hasBlockPoolId() != other.hasBlockPoolId()) return false;
if (hasBlockPoolId()) {
if (!getBlockPoolId()
.equals(other.getBlockPoolId())) return false;
}
if (hasWebAddress() != other.hasWebAddress()) return false;
if (hasWebAddress()) {
if (!getWebAddress()
.equals(other.getWebAddress())) return false;
}
if (hasRpcAddress() != other.hasRpcAddress()) return false;
if (hasRpcAddress()) {
if (!getRpcAddress()
.equals(other.getRpcAddress())) return false;
}
if (hasServiceAddress() != other.hasServiceAddress()) return false;
if (hasServiceAddress()) {
if (!getServiceAddress()
.equals(other.getServiceAddress())) return false;
}
if (hasLifelineAddress() != other.hasLifelineAddress()) return false;
if (hasLifelineAddress()) {
if (!getLifelineAddress()
.equals(other.getLifelineAddress())) return false;
}
if (hasState() != other.hasState()) return false;
if (hasState()) {
if (!getState()
.equals(other.getState())) return false;
}
if (hasIsSafeMode() != other.hasIsSafeMode()) return false;
if (hasIsSafeMode()) {
if (getIsSafeMode()
!= other.getIsSafeMode()) return false;
}
if (hasStats() != other.hasStats()) return false;
if (hasStats()) {
if (!getStats()
.equals(other.getStats())) return false;
}
if (hasWebScheme() != other.hasWebScheme()) return false;
if (hasWebScheme()) {
if (!getWebScheme()
.equals(other.getWebScheme())) return false;
}
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (hasDateCreated()) {
hash = (37 * hash) + DATECREATED_FIELD_NUMBER;
hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashLong(
getDateCreated());
}
if (hasDateModified()) {
hash = (37 * hash) + DATEMODIFIED_FIELD_NUMBER;
hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashLong(
getDateModified());
}
if (hasLastContact()) {
hash = (37 * hash) + LASTCONTACT_FIELD_NUMBER;
hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashLong(
getLastContact());
}
if (hasRouterId()) {
hash = (37 * hash) + ROUTERID_FIELD_NUMBER;
hash = (53 * hash) + getRouterId().hashCode();
}
if (hasNameserviceId()) {
hash = (37 * hash) + NAMESERVICEID_FIELD_NUMBER;
hash = (53 * hash) + getNameserviceId().hashCode();
}
if (hasNamenodeId()) {
hash = (37 * hash) + NAMENODEID_FIELD_NUMBER;
hash = (53 * hash) + getNamenodeId().hashCode();
}
if (hasClusterId()) {
hash = (37 * hash) + CLUSTERID_FIELD_NUMBER;
hash = (53 * hash) + getClusterId().hashCode();
}
if (hasBlockPoolId()) {
hash = (37 * hash) + BLOCKPOOLID_FIELD_NUMBER;
hash = (53 * hash) + getBlockPoolId().hashCode();
}
if (hasWebAddress()) {
hash = (37 * hash) + WEBADDRESS_FIELD_NUMBER;
hash = (53 * hash) + getWebAddress().hashCode();
}
if (hasRpcAddress()) {
hash = (37 * hash) + RPCADDRESS_FIELD_NUMBER;
hash = (53 * hash) + getRpcAddress().hashCode();
}
if (hasServiceAddress()) {
hash = (37 * hash) + SERVICEADDRESS_FIELD_NUMBER;
hash = (53 * hash) + getServiceAddress().hashCode();
}
if (hasLifelineAddress()) {
hash = (37 * hash) + LIFELINEADDRESS_FIELD_NUMBER;
hash = (53 * hash) + getLifelineAddress().hashCode();
}
if (hasState()) {
hash = (37 * hash) + STATE_FIELD_NUMBER;
hash = (53 * hash) + getState().hashCode();
}
if (hasIsSafeMode()) {
hash = (37 * hash) + ISSAFEMODE_FIELD_NUMBER;
hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashBoolean(
getIsSafeMode());
}
if (hasStats()) {
hash = (37 * hash) + STATS_FIELD_NUMBER;
hash = (53 * hash) + getStats().hashCode();
}
if (hasWebScheme()) {
hash = (37 * hash) + WEBSCHEME_FIELD_NUMBER;
hash = (53 * hash) + getWebScheme().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.NamenodeMembershipRecordProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.NamenodeMembershipRecordProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_NamenodeMembershipRecordProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_NamenodeMembershipRecordProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
if (org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
getStatsFieldBuilder();
}
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
dateCreated_ = 0L;
dateModified_ = 0L;
lastContact_ = 0L;
routerId_ = "";
nameserviceId_ = "";
namenodeId_ = "";
clusterId_ = "";
blockPoolId_ = "";
webAddress_ = "";
rpcAddress_ = "";
serviceAddress_ = "";
lifelineAddress_ = "";
state_ = "";
isSafeMode_ = false;
stats_ = null;
if (statsBuilder_ != null) {
statsBuilder_.dispose();
statsBuilder_ = null;
}
webScheme_ = "";
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_NamenodeMembershipRecordProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto(this);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartial0(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto result) {
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) != 0)) {
result.dateCreated_ = dateCreated_;
to_bitField0_ |= 0x00000001;
}
if (((from_bitField0_ & 0x00000002) != 0)) {
result.dateModified_ = dateModified_;
to_bitField0_ |= 0x00000002;
}
if (((from_bitField0_ & 0x00000004) != 0)) {
result.lastContact_ = lastContact_;
to_bitField0_ |= 0x00000004;
}
if (((from_bitField0_ & 0x00000008) != 0)) {
result.routerId_ = routerId_;
to_bitField0_ |= 0x00000008;
}
if (((from_bitField0_ & 0x00000010) != 0)) {
result.nameserviceId_ = nameserviceId_;
to_bitField0_ |= 0x00000010;
}
if (((from_bitField0_ & 0x00000020) != 0)) {
result.namenodeId_ = namenodeId_;
to_bitField0_ |= 0x00000020;
}
if (((from_bitField0_ & 0x00000040) != 0)) {
result.clusterId_ = clusterId_;
to_bitField0_ |= 0x00000040;
}
if (((from_bitField0_ & 0x00000080) != 0)) {
result.blockPoolId_ = blockPoolId_;
to_bitField0_ |= 0x00000080;
}
if (((from_bitField0_ & 0x00000100) != 0)) {
result.webAddress_ = webAddress_;
to_bitField0_ |= 0x00000100;
}
if (((from_bitField0_ & 0x00000200) != 0)) {
result.rpcAddress_ = rpcAddress_;
to_bitField0_ |= 0x00000200;
}
if (((from_bitField0_ & 0x00000400) != 0)) {
result.serviceAddress_ = serviceAddress_;
to_bitField0_ |= 0x00000400;
}
if (((from_bitField0_ & 0x00000800) != 0)) {
result.lifelineAddress_ = lifelineAddress_;
to_bitField0_ |= 0x00000800;
}
if (((from_bitField0_ & 0x00001000) != 0)) {
result.state_ = state_;
to_bitField0_ |= 0x00001000;
}
if (((from_bitField0_ & 0x00002000) != 0)) {
result.isSafeMode_ = isSafeMode_;
to_bitField0_ |= 0x00002000;
}
if (((from_bitField0_ & 0x00004000) != 0)) {
result.stats_ = statsBuilder_ == null
? stats_
: statsBuilder_.build();
to_bitField0_ |= 0x00004000;
}
if (((from_bitField0_ & 0x00008000) != 0)) {
result.webScheme_ = webScheme_;
to_bitField0_ |= 0x00008000;
}
result.bitField0_ |= to_bitField0_;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto.getDefaultInstance()) return this;
if (other.hasDateCreated()) {
setDateCreated(other.getDateCreated());
}
if (other.hasDateModified()) {
setDateModified(other.getDateModified());
}
if (other.hasLastContact()) {
setLastContact(other.getLastContact());
}
if (other.hasRouterId()) {
routerId_ = other.routerId_;
bitField0_ |= 0x00000008;
onChanged();
}
if (other.hasNameserviceId()) {
nameserviceId_ = other.nameserviceId_;
bitField0_ |= 0x00000010;
onChanged();
}
if (other.hasNamenodeId()) {
namenodeId_ = other.namenodeId_;
bitField0_ |= 0x00000020;
onChanged();
}
if (other.hasClusterId()) {
clusterId_ = other.clusterId_;
bitField0_ |= 0x00000040;
onChanged();
}
if (other.hasBlockPoolId()) {
blockPoolId_ = other.blockPoolId_;
bitField0_ |= 0x00000080;
onChanged();
}
if (other.hasWebAddress()) {
webAddress_ = other.webAddress_;
bitField0_ |= 0x00000100;
onChanged();
}
if (other.hasRpcAddress()) {
rpcAddress_ = other.rpcAddress_;
bitField0_ |= 0x00000200;
onChanged();
}
if (other.hasServiceAddress()) {
serviceAddress_ = other.serviceAddress_;
bitField0_ |= 0x00000400;
onChanged();
}
if (other.hasLifelineAddress()) {
lifelineAddress_ = other.lifelineAddress_;
bitField0_ |= 0x00000800;
onChanged();
}
if (other.hasState()) {
state_ = other.state_;
bitField0_ |= 0x00001000;
onChanged();
}
if (other.hasIsSafeMode()) {
setIsSafeMode(other.getIsSafeMode());
}
if (other.hasStats()) {
mergeStats(other.getStats());
}
if (other.hasWebScheme()) {
webScheme_ = other.webScheme_;
bitField0_ |= 0x00008000;
onChanged();
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 8: {
dateCreated_ = input.readUInt64();
bitField0_ |= 0x00000001;
break;
} // case 8
case 16: {
dateModified_ = input.readUInt64();
bitField0_ |= 0x00000002;
break;
} // case 16
case 24: {
lastContact_ = input.readUInt64();
bitField0_ |= 0x00000004;
break;
} // case 24
case 34: {
routerId_ = input.readBytes();
bitField0_ |= 0x00000008;
break;
} // case 34
case 42: {
nameserviceId_ = input.readBytes();
bitField0_ |= 0x00000010;
break;
} // case 42
case 50: {
namenodeId_ = input.readBytes();
bitField0_ |= 0x00000020;
break;
} // case 50
case 58: {
clusterId_ = input.readBytes();
bitField0_ |= 0x00000040;
break;
} // case 58
case 66: {
blockPoolId_ = input.readBytes();
bitField0_ |= 0x00000080;
break;
} // case 66
case 74: {
webAddress_ = input.readBytes();
bitField0_ |= 0x00000100;
break;
} // case 74
case 82: {
rpcAddress_ = input.readBytes();
bitField0_ |= 0x00000200;
break;
} // case 82
case 90: {
serviceAddress_ = input.readBytes();
bitField0_ |= 0x00000400;
break;
} // case 90
case 98: {
lifelineAddress_ = input.readBytes();
bitField0_ |= 0x00000800;
break;
} // case 98
case 106: {
state_ = input.readBytes();
bitField0_ |= 0x00001000;
break;
} // case 106
case 112: {
isSafeMode_ = input.readBool();
bitField0_ |= 0x00002000;
break;
} // case 112
case 122: {
input.readMessage(
getStatsFieldBuilder().getBuilder(),
extensionRegistry);
bitField0_ |= 0x00004000;
break;
} // case 122
case 130: {
webScheme_ = input.readBytes();
bitField0_ |= 0x00008000;
break;
} // case 130
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private long dateCreated_ ;
/**
* <code>optional uint64 dateCreated = 1;</code>
* @return Whether the dateCreated field is set.
*/
@java.lang.Override
public boolean hasDateCreated() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional uint64 dateCreated = 1;</code>
* @return The dateCreated.
*/
@java.lang.Override
public long getDateCreated() {
return dateCreated_;
}
/**
* <code>optional uint64 dateCreated = 1;</code>
* @param value The dateCreated to set.
* @return This builder for chaining.
*/
public Builder setDateCreated(long value) {
dateCreated_ = value;
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>optional uint64 dateCreated = 1;</code>
* @return This builder for chaining.
*/
public Builder clearDateCreated() {
bitField0_ = (bitField0_ & ~0x00000001);
dateCreated_ = 0L;
onChanged();
return this;
}
private long dateModified_ ;
/**
* <code>optional uint64 dateModified = 2;</code>
* @return Whether the dateModified field is set.
*/
@java.lang.Override
public boolean hasDateModified() {
return ((bitField0_ & 0x00000002) != 0);
}
/**
* <code>optional uint64 dateModified = 2;</code>
* @return The dateModified.
*/
@java.lang.Override
public long getDateModified() {
return dateModified_;
}
/**
* <code>optional uint64 dateModified = 2;</code>
* @param value The dateModified to set.
* @return This builder for chaining.
*/
public Builder setDateModified(long value) {
dateModified_ = value;
bitField0_ |= 0x00000002;
onChanged();
return this;
}
/**
* <code>optional uint64 dateModified = 2;</code>
* @return This builder for chaining.
*/
public Builder clearDateModified() {
bitField0_ = (bitField0_ & ~0x00000002);
dateModified_ = 0L;
onChanged();
return this;
}
private long lastContact_ ;
/**
* <code>optional uint64 lastContact = 3;</code>
* @return Whether the lastContact field is set.
*/
@java.lang.Override
public boolean hasLastContact() {
return ((bitField0_ & 0x00000004) != 0);
}
/**
* <code>optional uint64 lastContact = 3;</code>
* @return The lastContact.
*/
@java.lang.Override
public long getLastContact() {
return lastContact_;
}
/**
* <code>optional uint64 lastContact = 3;</code>
* @param value The lastContact to set.
* @return This builder for chaining.
*/
public Builder setLastContact(long value) {
lastContact_ = value;
bitField0_ |= 0x00000004;
onChanged();
return this;
}
/**
* <code>optional uint64 lastContact = 3;</code>
* @return This builder for chaining.
*/
public Builder clearLastContact() {
bitField0_ = (bitField0_ & ~0x00000004);
lastContact_ = 0L;
onChanged();
return this;
}
private java.lang.Object routerId_ = "";
/**
* <code>optional string routerId = 4;</code>
* @return Whether the routerId field is set.
*/
public boolean hasRouterId() {
return ((bitField0_ & 0x00000008) != 0);
}
/**
* <code>optional string routerId = 4;</code>
* @return The routerId.
*/
public java.lang.String getRouterId() {
java.lang.Object ref = routerId_;
if (!(ref instanceof java.lang.String)) {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
routerId_ = s;
}
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* <code>optional string routerId = 4;</code>
* @return The bytes for routerId.
*/
public org.apache.hadoop.thirdparty.protobuf.ByteString
getRouterIdBytes() {
java.lang.Object ref = routerId_;
if (ref instanceof String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
routerId_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
/**
* <code>optional string routerId = 4;</code>
* @param value The routerId to set.
* @return This builder for chaining.
*/
public Builder setRouterId(
java.lang.String value) {
if (value == null) { throw new NullPointerException(); }
routerId_ = value;
bitField0_ |= 0x00000008;
onChanged();
return this;
}
/**
* <code>optional string routerId = 4;</code>
* @return This builder for chaining.
*/
public Builder clearRouterId() {
routerId_ = getDefaultInstance().getRouterId();
bitField0_ = (bitField0_ & ~0x00000008);
onChanged();
return this;
}
/**
* <code>optional string routerId = 4;</code>
* @param value The bytes for routerId to set.
* @return This builder for chaining.
*/
public Builder setRouterIdBytes(
org.apache.hadoop.thirdparty.protobuf.ByteString value) {
if (value == null) { throw new NullPointerException(); }
routerId_ = value;
bitField0_ |= 0x00000008;
onChanged();
return this;
}
private java.lang.Object nameserviceId_ = "";
/**
* <code>optional string nameserviceId = 5;</code>
* @return Whether the nameserviceId field is set.
*/
public boolean hasNameserviceId() {
return ((bitField0_ & 0x00000010) != 0);
}
/**
* <code>optional string nameserviceId = 5;</code>
* @return The nameserviceId.
*/
public java.lang.String getNameserviceId() {
java.lang.Object ref = nameserviceId_;
if (!(ref instanceof java.lang.String)) {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
nameserviceId_ = s;
}
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* <code>optional string nameserviceId = 5;</code>
* @return The bytes for nameserviceId.
*/
public org.apache.hadoop.thirdparty.protobuf.ByteString
getNameserviceIdBytes() {
java.lang.Object ref = nameserviceId_;
if (ref instanceof String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
nameserviceId_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
/**
* <code>optional string nameserviceId = 5;</code>
* @param value The nameserviceId to set.
* @return This builder for chaining.
*/
public Builder setNameserviceId(
java.lang.String value) {
if (value == null) { throw new NullPointerException(); }
nameserviceId_ = value;
bitField0_ |= 0x00000010;
onChanged();
return this;
}
/**
* <code>optional string nameserviceId = 5;</code>
* @return This builder for chaining.
*/
public Builder clearNameserviceId() {
nameserviceId_ = getDefaultInstance().getNameserviceId();
bitField0_ = (bitField0_ & ~0x00000010);
onChanged();
return this;
}
/**
* <code>optional string nameserviceId = 5;</code>
* @param value The bytes for nameserviceId to set.
* @return This builder for chaining.
*/
public Builder setNameserviceIdBytes(
org.apache.hadoop.thirdparty.protobuf.ByteString value) {
if (value == null) { throw new NullPointerException(); }
nameserviceId_ = value;
bitField0_ |= 0x00000010;
onChanged();
return this;
}
private java.lang.Object namenodeId_ = "";
/**
* <code>optional string namenodeId = 6;</code>
* @return Whether the namenodeId field is set.
*/
public boolean hasNamenodeId() {
return ((bitField0_ & 0x00000020) != 0);
}
/**
* <code>optional string namenodeId = 6;</code>
* @return The namenodeId.
*/
public java.lang.String getNamenodeId() {
java.lang.Object ref = namenodeId_;
if (!(ref instanceof java.lang.String)) {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
namenodeId_ = s;
}
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* <code>optional string namenodeId = 6;</code>
* @return The bytes for namenodeId.
*/
public org.apache.hadoop.thirdparty.protobuf.ByteString
getNamenodeIdBytes() {
java.lang.Object ref = namenodeId_;
if (ref instanceof String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
namenodeId_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
/**
* <code>optional string namenodeId = 6;</code>
* @param value The namenodeId to set.
* @return This builder for chaining.
*/
public Builder setNamenodeId(
java.lang.String value) {
if (value == null) { throw new NullPointerException(); }
namenodeId_ = value;
bitField0_ |= 0x00000020;
onChanged();
return this;
}
/**
* <code>optional string namenodeId = 6;</code>
* @return This builder for chaining.
*/
public Builder clearNamenodeId() {
namenodeId_ = getDefaultInstance().getNamenodeId();
bitField0_ = (bitField0_ & ~0x00000020);
onChanged();
return this;
}
/**
* <code>optional string namenodeId = 6;</code>
* @param value The bytes for namenodeId to set.
* @return This builder for chaining.
*/
public Builder setNamenodeIdBytes(
org.apache.hadoop.thirdparty.protobuf.ByteString value) {
if (value == null) { throw new NullPointerException(); }
namenodeId_ = value;
bitField0_ |= 0x00000020;
onChanged();
return this;
}
private java.lang.Object clusterId_ = "";
/**
* <code>optional string clusterId = 7;</code>
* @return Whether the clusterId field is set.
*/
public boolean hasClusterId() {
return ((bitField0_ & 0x00000040) != 0);
}
/**
* <code>optional string clusterId = 7;</code>
* @return The clusterId.
*/
public java.lang.String getClusterId() {
java.lang.Object ref = clusterId_;
if (!(ref instanceof java.lang.String)) {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
clusterId_ = s;
}
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* <code>optional string clusterId = 7;</code>
* @return The bytes for clusterId.
*/
public org.apache.hadoop.thirdparty.protobuf.ByteString
getClusterIdBytes() {
java.lang.Object ref = clusterId_;
if (ref instanceof String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
clusterId_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
/**
* <code>optional string clusterId = 7;</code>
* @param value The clusterId to set.
* @return This builder for chaining.
*/
public Builder setClusterId(
java.lang.String value) {
if (value == null) { throw new NullPointerException(); }
clusterId_ = value;
bitField0_ |= 0x00000040;
onChanged();
return this;
}
/**
* <code>optional string clusterId = 7;</code>
* @return This builder for chaining.
*/
public Builder clearClusterId() {
clusterId_ = getDefaultInstance().getClusterId();
bitField0_ = (bitField0_ & ~0x00000040);
onChanged();
return this;
}
/**
* <code>optional string clusterId = 7;</code>
* @param value The bytes for clusterId to set.
* @return This builder for chaining.
*/
public Builder setClusterIdBytes(
org.apache.hadoop.thirdparty.protobuf.ByteString value) {
if (value == null) { throw new NullPointerException(); }
clusterId_ = value;
bitField0_ |= 0x00000040;
onChanged();
return this;
}
private java.lang.Object blockPoolId_ = "";
/**
* <code>optional string blockPoolId = 8;</code>
* @return Whether the blockPoolId field is set.
*/
public boolean hasBlockPoolId() {
return ((bitField0_ & 0x00000080) != 0);
}
/**
* <code>optional string blockPoolId = 8;</code>
* @return The blockPoolId.
*/
public java.lang.String getBlockPoolId() {
java.lang.Object ref = blockPoolId_;
if (!(ref instanceof java.lang.String)) {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
blockPoolId_ = s;
}
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* <code>optional string blockPoolId = 8;</code>
* @return The bytes for blockPoolId.
*/
public org.apache.hadoop.thirdparty.protobuf.ByteString
getBlockPoolIdBytes() {
java.lang.Object ref = blockPoolId_;
if (ref instanceof String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
blockPoolId_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
/**
* <code>optional string blockPoolId = 8;</code>
* @param value The blockPoolId to set.
* @return This builder for chaining.
*/
public Builder setBlockPoolId(
java.lang.String value) {
if (value == null) { throw new NullPointerException(); }
blockPoolId_ = value;
bitField0_ |= 0x00000080;
onChanged();
return this;
}
/**
* <code>optional string blockPoolId = 8;</code>
* @return This builder for chaining.
*/
public Builder clearBlockPoolId() {
blockPoolId_ = getDefaultInstance().getBlockPoolId();
bitField0_ = (bitField0_ & ~0x00000080);
onChanged();
return this;
}
/**
* <code>optional string blockPoolId = 8;</code>
* @param value The bytes for blockPoolId to set.
* @return This builder for chaining.
*/
public Builder setBlockPoolIdBytes(
org.apache.hadoop.thirdparty.protobuf.ByteString value) {
if (value == null) { throw new NullPointerException(); }
blockPoolId_ = value;
bitField0_ |= 0x00000080;
onChanged();
return this;
}
private java.lang.Object webAddress_ = "";
/**
* <code>optional string webAddress = 9;</code>
* @return Whether the webAddress field is set.
*/
public boolean hasWebAddress() {
return ((bitField0_ & 0x00000100) != 0);
}
/**
* <code>optional string webAddress = 9;</code>
* @return The webAddress.
*/
public java.lang.String getWebAddress() {
java.lang.Object ref = webAddress_;
if (!(ref instanceof java.lang.String)) {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
webAddress_ = s;
}
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* <code>optional string webAddress = 9;</code>
* @return The bytes for webAddress.
*/
public org.apache.hadoop.thirdparty.protobuf.ByteString
getWebAddressBytes() {
java.lang.Object ref = webAddress_;
if (ref instanceof String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
webAddress_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
/**
* <code>optional string webAddress = 9;</code>
* @param value The webAddress to set.
* @return This builder for chaining.
*/
public Builder setWebAddress(
java.lang.String value) {
if (value == null) { throw new NullPointerException(); }
webAddress_ = value;
bitField0_ |= 0x00000100;
onChanged();
return this;
}
/**
* <code>optional string webAddress = 9;</code>
* @return This builder for chaining.
*/
public Builder clearWebAddress() {
webAddress_ = getDefaultInstance().getWebAddress();
bitField0_ = (bitField0_ & ~0x00000100);
onChanged();
return this;
}
/**
* <code>optional string webAddress = 9;</code>
* @param value The bytes for webAddress to set.
* @return This builder for chaining.
*/
public Builder setWebAddressBytes(
org.apache.hadoop.thirdparty.protobuf.ByteString value) {
if (value == null) { throw new NullPointerException(); }
webAddress_ = value;
bitField0_ |= 0x00000100;
onChanged();
return this;
}
private java.lang.Object rpcAddress_ = "";
/**
* <code>optional string rpcAddress = 10;</code>
* @return Whether the rpcAddress field is set.
*/
public boolean hasRpcAddress() {
return ((bitField0_ & 0x00000200) != 0);
}
/**
* <code>optional string rpcAddress = 10;</code>
* @return The rpcAddress.
*/
public java.lang.String getRpcAddress() {
java.lang.Object ref = rpcAddress_;
if (!(ref instanceof java.lang.String)) {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
rpcAddress_ = s;
}
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* <code>optional string rpcAddress = 10;</code>
* @return The bytes for rpcAddress.
*/
public org.apache.hadoop.thirdparty.protobuf.ByteString
getRpcAddressBytes() {
java.lang.Object ref = rpcAddress_;
if (ref instanceof String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
rpcAddress_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
/**
* <code>optional string rpcAddress = 10;</code>
* @param value The rpcAddress to set.
* @return This builder for chaining.
*/
public Builder setRpcAddress(
java.lang.String value) {
if (value == null) { throw new NullPointerException(); }
rpcAddress_ = value;
bitField0_ |= 0x00000200;
onChanged();
return this;
}
/**
* <code>optional string rpcAddress = 10;</code>
* @return This builder for chaining.
*/
public Builder clearRpcAddress() {
rpcAddress_ = getDefaultInstance().getRpcAddress();
bitField0_ = (bitField0_ & ~0x00000200);
onChanged();
return this;
}
/**
* <code>optional string rpcAddress = 10;</code>
* @param value The bytes for rpcAddress to set.
* @return This builder for chaining.
*/
public Builder setRpcAddressBytes(
org.apache.hadoop.thirdparty.protobuf.ByteString value) {
if (value == null) { throw new NullPointerException(); }
rpcAddress_ = value;
bitField0_ |= 0x00000200;
onChanged();
return this;
}
private java.lang.Object serviceAddress_ = "";
/**
* <code>optional string serviceAddress = 11;</code>
* @return Whether the serviceAddress field is set.
*/
public boolean hasServiceAddress() {
return ((bitField0_ & 0x00000400) != 0);
}
/**
* <code>optional string serviceAddress = 11;</code>
* @return The serviceAddress.
*/
public java.lang.String getServiceAddress() {
java.lang.Object ref = serviceAddress_;
if (!(ref instanceof java.lang.String)) {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
serviceAddress_ = s;
}
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* <code>optional string serviceAddress = 11;</code>
* @return The bytes for serviceAddress.
*/
public org.apache.hadoop.thirdparty.protobuf.ByteString
getServiceAddressBytes() {
java.lang.Object ref = serviceAddress_;
if (ref instanceof String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
serviceAddress_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
/**
* <code>optional string serviceAddress = 11;</code>
* @param value The serviceAddress to set.
* @return This builder for chaining.
*/
public Builder setServiceAddress(
java.lang.String value) {
if (value == null) { throw new NullPointerException(); }
serviceAddress_ = value;
bitField0_ |= 0x00000400;
onChanged();
return this;
}
/**
* <code>optional string serviceAddress = 11;</code>
* @return This builder for chaining.
*/
public Builder clearServiceAddress() {
serviceAddress_ = getDefaultInstance().getServiceAddress();
bitField0_ = (bitField0_ & ~0x00000400);
onChanged();
return this;
}
/**
* <code>optional string serviceAddress = 11;</code>
* @param value The bytes for serviceAddress to set.
* @return This builder for chaining.
*/
public Builder setServiceAddressBytes(
org.apache.hadoop.thirdparty.protobuf.ByteString value) {
if (value == null) { throw new NullPointerException(); }
serviceAddress_ = value;
bitField0_ |= 0x00000400;
onChanged();
return this;
}
private java.lang.Object lifelineAddress_ = "";
/**
* <code>optional string lifelineAddress = 12;</code>
* @return Whether the lifelineAddress field is set.
*/
public boolean hasLifelineAddress() {
return ((bitField0_ & 0x00000800) != 0);
}
/**
* <code>optional string lifelineAddress = 12;</code>
* @return The lifelineAddress.
*/
public java.lang.String getLifelineAddress() {
java.lang.Object ref = lifelineAddress_;
if (!(ref instanceof java.lang.String)) {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
lifelineAddress_ = s;
}
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* <code>optional string lifelineAddress = 12;</code>
* @return The bytes for lifelineAddress.
*/
public org.apache.hadoop.thirdparty.protobuf.ByteString
getLifelineAddressBytes() {
java.lang.Object ref = lifelineAddress_;
if (ref instanceof String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
lifelineAddress_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
/**
* <code>optional string lifelineAddress = 12;</code>
* @param value The lifelineAddress to set.
* @return This builder for chaining.
*/
public Builder setLifelineAddress(
java.lang.String value) {
if (value == null) { throw new NullPointerException(); }
lifelineAddress_ = value;
bitField0_ |= 0x00000800;
onChanged();
return this;
}
/**
* <code>optional string lifelineAddress = 12;</code>
* @return This builder for chaining.
*/
public Builder clearLifelineAddress() {
lifelineAddress_ = getDefaultInstance().getLifelineAddress();
bitField0_ = (bitField0_ & ~0x00000800);
onChanged();
return this;
}
/**
* <code>optional string lifelineAddress = 12;</code>
* @param value The bytes for lifelineAddress to set.
* @return This builder for chaining.
*/
public Builder setLifelineAddressBytes(
org.apache.hadoop.thirdparty.protobuf.ByteString value) {
if (value == null) { throw new NullPointerException(); }
lifelineAddress_ = value;
bitField0_ |= 0x00000800;
onChanged();
return this;
}
private java.lang.Object state_ = "";
/**
* <code>optional string state = 13;</code>
* @return Whether the state field is set.
*/
public boolean hasState() {
return ((bitField0_ & 0x00001000) != 0);
}
/**
* <code>optional string state = 13;</code>
* @return The state.
*/
public java.lang.String getState() {
java.lang.Object ref = state_;
if (!(ref instanceof java.lang.String)) {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
state_ = s;
}
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* <code>optional string state = 13;</code>
* @return The bytes for state.
*/
public org.apache.hadoop.thirdparty.protobuf.ByteString
getStateBytes() {
java.lang.Object ref = state_;
if (ref instanceof String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
state_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
/**
* <code>optional string state = 13;</code>
* @param value The state to set.
* @return This builder for chaining.
*/
public Builder setState(
java.lang.String value) {
if (value == null) { throw new NullPointerException(); }
state_ = value;
bitField0_ |= 0x00001000;
onChanged();
return this;
}
/**
* <code>optional string state = 13;</code>
* @return This builder for chaining.
*/
public Builder clearState() {
state_ = getDefaultInstance().getState();
bitField0_ = (bitField0_ & ~0x00001000);
onChanged();
return this;
}
/**
* <code>optional string state = 13;</code>
* @param value The bytes for state to set.
* @return This builder for chaining.
*/
public Builder setStateBytes(
org.apache.hadoop.thirdparty.protobuf.ByteString value) {
if (value == null) { throw new NullPointerException(); }
state_ = value;
bitField0_ |= 0x00001000;
onChanged();
return this;
}
private boolean isSafeMode_ ;
/**
* <code>optional bool isSafeMode = 14;</code>
* @return Whether the isSafeMode field is set.
*/
@java.lang.Override
public boolean hasIsSafeMode() {
return ((bitField0_ & 0x00002000) != 0);
}
/**
* <code>optional bool isSafeMode = 14;</code>
* @return The isSafeMode.
*/
@java.lang.Override
public boolean getIsSafeMode() {
return isSafeMode_;
}
/**
* <code>optional bool isSafeMode = 14;</code>
* @param value The isSafeMode to set.
* @return This builder for chaining.
*/
public Builder setIsSafeMode(boolean value) {
isSafeMode_ = value;
bitField0_ |= 0x00002000;
onChanged();
return this;
}
/**
* <code>optional bool isSafeMode = 14;</code>
* @return This builder for chaining.
*/
public Builder clearIsSafeMode() {
bitField0_ = (bitField0_ & ~0x00002000);
isSafeMode_ = false;
onChanged();
return this;
}
private org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto stats_;
private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto.Builder, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProtoOrBuilder> statsBuilder_;
/**
* <code>optional .hadoop.hdfs.NamenodeMembershipStatsRecordProto stats = 15;</code>
* @return Whether the stats field is set.
*/
public boolean hasStats() {
return ((bitField0_ & 0x00004000) != 0);
}
/**
* <code>optional .hadoop.hdfs.NamenodeMembershipStatsRecordProto stats = 15;</code>
* @return The stats.
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto getStats() {
if (statsBuilder_ == null) {
return stats_ == null ? org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto.getDefaultInstance() : stats_;
} else {
return statsBuilder_.getMessage();
}
}
/**
* <code>optional .hadoop.hdfs.NamenodeMembershipStatsRecordProto stats = 15;</code>
*/
public Builder setStats(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto value) {
if (statsBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
stats_ = value;
} else {
statsBuilder_.setMessage(value);
}
bitField0_ |= 0x00004000;
onChanged();
return this;
}
/**
* <code>optional .hadoop.hdfs.NamenodeMembershipStatsRecordProto stats = 15;</code>
*/
public Builder setStats(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto.Builder builderForValue) {
if (statsBuilder_ == null) {
stats_ = builderForValue.build();
} else {
statsBuilder_.setMessage(builderForValue.build());
}
bitField0_ |= 0x00004000;
onChanged();
return this;
}
/**
* <code>optional .hadoop.hdfs.NamenodeMembershipStatsRecordProto stats = 15;</code>
*/
public Builder mergeStats(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto value) {
if (statsBuilder_ == null) {
if (((bitField0_ & 0x00004000) != 0) &&
stats_ != null &&
stats_ != org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto.getDefaultInstance()) {
getStatsBuilder().mergeFrom(value);
} else {
stats_ = value;
}
} else {
statsBuilder_.mergeFrom(value);
}
if (stats_ != null) {
bitField0_ |= 0x00004000;
onChanged();
}
return this;
}
/**
* <code>optional .hadoop.hdfs.NamenodeMembershipStatsRecordProto stats = 15;</code>
*/
public Builder clearStats() {
bitField0_ = (bitField0_ & ~0x00004000);
stats_ = null;
if (statsBuilder_ != null) {
statsBuilder_.dispose();
statsBuilder_ = null;
}
onChanged();
return this;
}
/**
* <code>optional .hadoop.hdfs.NamenodeMembershipStatsRecordProto stats = 15;</code>
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto.Builder getStatsBuilder() {
bitField0_ |= 0x00004000;
onChanged();
return getStatsFieldBuilder().getBuilder();
}
/**
* <code>optional .hadoop.hdfs.NamenodeMembershipStatsRecordProto stats = 15;</code>
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProtoOrBuilder getStatsOrBuilder() {
if (statsBuilder_ != null) {
return statsBuilder_.getMessageOrBuilder();
} else {
return stats_ == null ?
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto.getDefaultInstance() : stats_;
}
}
/**
* <code>optional .hadoop.hdfs.NamenodeMembershipStatsRecordProto stats = 15;</code>
*/
private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto.Builder, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProtoOrBuilder>
getStatsFieldBuilder() {
if (statsBuilder_ == null) {
statsBuilder_ = new org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProto.Builder, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipStatsRecordProtoOrBuilder>(
getStats(),
getParentForChildren(),
isClean());
stats_ = null;
}
return statsBuilder_;
}
private java.lang.Object webScheme_ = "";
/**
* <code>optional string webScheme = 16;</code>
* @return Whether the webScheme field is set.
*/
public boolean hasWebScheme() {
return ((bitField0_ & 0x00008000) != 0);
}
/**
* <code>optional string webScheme = 16;</code>
* @return The webScheme.
*/
public java.lang.String getWebScheme() {
java.lang.Object ref = webScheme_;
if (!(ref instanceof java.lang.String)) {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
webScheme_ = s;
}
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* <code>optional string webScheme = 16;</code>
* @return The bytes for webScheme.
*/
public org.apache.hadoop.thirdparty.protobuf.ByteString
getWebSchemeBytes() {
java.lang.Object ref = webScheme_;
if (ref instanceof String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
webScheme_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
/**
* <code>optional string webScheme = 16;</code>
* @param value The webScheme to set.
* @return This builder for chaining.
*/
public Builder setWebScheme(
java.lang.String value) {
if (value == null) { throw new NullPointerException(); }
webScheme_ = value;
bitField0_ |= 0x00008000;
onChanged();
return this;
}
/**
* <code>optional string webScheme = 16;</code>
* @return This builder for chaining.
*/
public Builder clearWebScheme() {
webScheme_ = getDefaultInstance().getWebScheme();
bitField0_ = (bitField0_ & ~0x00008000);
onChanged();
return this;
}
/**
* <code>optional string webScheme = 16;</code>
* @param value The bytes for webScheme to set.
* @return This builder for chaining.
*/
public Builder setWebSchemeBytes(
org.apache.hadoop.thirdparty.protobuf.ByteString value) {
if (value == null) { throw new NullPointerException(); }
webScheme_ = value;
bitField0_ |= 0x00008000;
onChanged();
return this;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.NamenodeMembershipRecordProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.NamenodeMembershipRecordProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<NamenodeMembershipRecordProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<NamenodeMembershipRecordProto>() {
@java.lang.Override
public NamenodeMembershipRecordProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<NamenodeMembershipRecordProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<NamenodeMembershipRecordProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface FederationNamespaceInfoProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.FederationNamespaceInfoProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* <code>optional string blockPoolId = 1;</code>
* @return Whether the blockPoolId field is set.
*/
boolean hasBlockPoolId();
/**
* <code>optional string blockPoolId = 1;</code>
* @return The blockPoolId.
*/
java.lang.String getBlockPoolId();
/**
* <code>optional string blockPoolId = 1;</code>
* @return The bytes for blockPoolId.
*/
org.apache.hadoop.thirdparty.protobuf.ByteString
getBlockPoolIdBytes();
/**
* <code>optional string clusterId = 2;</code>
* @return Whether the clusterId field is set.
*/
boolean hasClusterId();
/**
* <code>optional string clusterId = 2;</code>
* @return The clusterId.
*/
java.lang.String getClusterId();
/**
* <code>optional string clusterId = 2;</code>
* @return The bytes for clusterId.
*/
org.apache.hadoop.thirdparty.protobuf.ByteString
getClusterIdBytes();
/**
* <code>optional string nameserviceId = 3;</code>
* @return Whether the nameserviceId field is set.
*/
boolean hasNameserviceId();
/**
* <code>optional string nameserviceId = 3;</code>
* @return The nameserviceId.
*/
java.lang.String getNameserviceId();
/**
* <code>optional string nameserviceId = 3;</code>
* @return The bytes for nameserviceId.
*/
org.apache.hadoop.thirdparty.protobuf.ByteString
getNameserviceIdBytes();
}
/**
* Protobuf type {@code hadoop.hdfs.FederationNamespaceInfoProto}
*/
public static final class FederationNamespaceInfoProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.FederationNamespaceInfoProto)
FederationNamespaceInfoProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use FederationNamespaceInfoProto.newBuilder() to construct.
private FederationNamespaceInfoProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private FederationNamespaceInfoProto() {
blockPoolId_ = "";
clusterId_ = "";
nameserviceId_ = "";
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new FederationNamespaceInfoProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_FederationNamespaceInfoProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_FederationNamespaceInfoProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto.Builder.class);
}
private int bitField0_;
public static final int BLOCKPOOLID_FIELD_NUMBER = 1;
@SuppressWarnings("serial")
private volatile java.lang.Object blockPoolId_ = "";
/**
* <code>optional string blockPoolId = 1;</code>
* @return Whether the blockPoolId field is set.
*/
@java.lang.Override
public boolean hasBlockPoolId() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional string blockPoolId = 1;</code>
* @return The blockPoolId.
*/
@java.lang.Override
public java.lang.String getBlockPoolId() {
java.lang.Object ref = blockPoolId_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
blockPoolId_ = s;
}
return s;
}
}
/**
* <code>optional string blockPoolId = 1;</code>
* @return The bytes for blockPoolId.
*/
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.ByteString
getBlockPoolIdBytes() {
java.lang.Object ref = blockPoolId_;
if (ref instanceof java.lang.String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
blockPoolId_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
public static final int CLUSTERID_FIELD_NUMBER = 2;
@SuppressWarnings("serial")
private volatile java.lang.Object clusterId_ = "";
/**
* <code>optional string clusterId = 2;</code>
* @return Whether the clusterId field is set.
*/
@java.lang.Override
public boolean hasClusterId() {
return ((bitField0_ & 0x00000002) != 0);
}
/**
* <code>optional string clusterId = 2;</code>
* @return The clusterId.
*/
@java.lang.Override
public java.lang.String getClusterId() {
java.lang.Object ref = clusterId_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
clusterId_ = s;
}
return s;
}
}
/**
* <code>optional string clusterId = 2;</code>
* @return The bytes for clusterId.
*/
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.ByteString
getClusterIdBytes() {
java.lang.Object ref = clusterId_;
if (ref instanceof java.lang.String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
clusterId_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
public static final int NAMESERVICEID_FIELD_NUMBER = 3;
@SuppressWarnings("serial")
private volatile java.lang.Object nameserviceId_ = "";
/**
* <code>optional string nameserviceId = 3;</code>
* @return Whether the nameserviceId field is set.
*/
@java.lang.Override
public boolean hasNameserviceId() {
return ((bitField0_ & 0x00000004) != 0);
}
/**
* <code>optional string nameserviceId = 3;</code>
* @return The nameserviceId.
*/
@java.lang.Override
public java.lang.String getNameserviceId() {
java.lang.Object ref = nameserviceId_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
nameserviceId_ = s;
}
return s;
}
}
/**
* <code>optional string nameserviceId = 3;</code>
* @return The bytes for nameserviceId.
*/
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.ByteString
getNameserviceIdBytes() {
java.lang.Object ref = nameserviceId_;
if (ref instanceof java.lang.String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
nameserviceId_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) != 0)) {
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 1, blockPoolId_);
}
if (((bitField0_ & 0x00000002) != 0)) {
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 2, clusterId_);
}
if (((bitField0_ & 0x00000004) != 0)) {
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 3, nameserviceId_);
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(1, blockPoolId_);
}
if (((bitField0_ & 0x00000002) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(2, clusterId_);
}
if (((bitField0_ & 0x00000004) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(3, nameserviceId_);
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto) obj;
if (hasBlockPoolId() != other.hasBlockPoolId()) return false;
if (hasBlockPoolId()) {
if (!getBlockPoolId()
.equals(other.getBlockPoolId())) return false;
}
if (hasClusterId() != other.hasClusterId()) return false;
if (hasClusterId()) {
if (!getClusterId()
.equals(other.getClusterId())) return false;
}
if (hasNameserviceId() != other.hasNameserviceId()) return false;
if (hasNameserviceId()) {
if (!getNameserviceId()
.equals(other.getNameserviceId())) return false;
}
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (hasBlockPoolId()) {
hash = (37 * hash) + BLOCKPOOLID_FIELD_NUMBER;
hash = (53 * hash) + getBlockPoolId().hashCode();
}
if (hasClusterId()) {
hash = (37 * hash) + CLUSTERID_FIELD_NUMBER;
hash = (53 * hash) + getClusterId().hashCode();
}
if (hasNameserviceId()) {
hash = (37 * hash) + NAMESERVICEID_FIELD_NUMBER;
hash = (53 * hash) + getNameserviceId().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.FederationNamespaceInfoProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.FederationNamespaceInfoProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_FederationNamespaceInfoProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_FederationNamespaceInfoProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
blockPoolId_ = "";
clusterId_ = "";
nameserviceId_ = "";
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_FederationNamespaceInfoProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto(this);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartial0(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto result) {
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) != 0)) {
result.blockPoolId_ = blockPoolId_;
to_bitField0_ |= 0x00000001;
}
if (((from_bitField0_ & 0x00000002) != 0)) {
result.clusterId_ = clusterId_;
to_bitField0_ |= 0x00000002;
}
if (((from_bitField0_ & 0x00000004) != 0)) {
result.nameserviceId_ = nameserviceId_;
to_bitField0_ |= 0x00000004;
}
result.bitField0_ |= to_bitField0_;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto.getDefaultInstance()) return this;
if (other.hasBlockPoolId()) {
blockPoolId_ = other.blockPoolId_;
bitField0_ |= 0x00000001;
onChanged();
}
if (other.hasClusterId()) {
clusterId_ = other.clusterId_;
bitField0_ |= 0x00000002;
onChanged();
}
if (other.hasNameserviceId()) {
nameserviceId_ = other.nameserviceId_;
bitField0_ |= 0x00000004;
onChanged();
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 10: {
blockPoolId_ = input.readBytes();
bitField0_ |= 0x00000001;
break;
} // case 10
case 18: {
clusterId_ = input.readBytes();
bitField0_ |= 0x00000002;
break;
} // case 18
case 26: {
nameserviceId_ = input.readBytes();
bitField0_ |= 0x00000004;
break;
} // case 26
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private java.lang.Object blockPoolId_ = "";
/**
* <code>optional string blockPoolId = 1;</code>
* @return Whether the blockPoolId field is set.
*/
public boolean hasBlockPoolId() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional string blockPoolId = 1;</code>
* @return The blockPoolId.
*/
public java.lang.String getBlockPoolId() {
java.lang.Object ref = blockPoolId_;
if (!(ref instanceof java.lang.String)) {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
blockPoolId_ = s;
}
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* <code>optional string blockPoolId = 1;</code>
* @return The bytes for blockPoolId.
*/
public org.apache.hadoop.thirdparty.protobuf.ByteString
getBlockPoolIdBytes() {
java.lang.Object ref = blockPoolId_;
if (ref instanceof String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
blockPoolId_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
/**
* <code>optional string blockPoolId = 1;</code>
* @param value The blockPoolId to set.
* @return This builder for chaining.
*/
public Builder setBlockPoolId(
java.lang.String value) {
if (value == null) { throw new NullPointerException(); }
blockPoolId_ = value;
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>optional string blockPoolId = 1;</code>
* @return This builder for chaining.
*/
public Builder clearBlockPoolId() {
blockPoolId_ = getDefaultInstance().getBlockPoolId();
bitField0_ = (bitField0_ & ~0x00000001);
onChanged();
return this;
}
/**
* <code>optional string blockPoolId = 1;</code>
* @param value The bytes for blockPoolId to set.
* @return This builder for chaining.
*/
public Builder setBlockPoolIdBytes(
org.apache.hadoop.thirdparty.protobuf.ByteString value) {
if (value == null) { throw new NullPointerException(); }
blockPoolId_ = value;
bitField0_ |= 0x00000001;
onChanged();
return this;
}
private java.lang.Object clusterId_ = "";
/**
* <code>optional string clusterId = 2;</code>
* @return Whether the clusterId field is set.
*/
public boolean hasClusterId() {
return ((bitField0_ & 0x00000002) != 0);
}
/**
* <code>optional string clusterId = 2;</code>
* @return The clusterId.
*/
public java.lang.String getClusterId() {
java.lang.Object ref = clusterId_;
if (!(ref instanceof java.lang.String)) {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
clusterId_ = s;
}
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* <code>optional string clusterId = 2;</code>
* @return The bytes for clusterId.
*/
public org.apache.hadoop.thirdparty.protobuf.ByteString
getClusterIdBytes() {
java.lang.Object ref = clusterId_;
if (ref instanceof String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
clusterId_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
/**
* <code>optional string clusterId = 2;</code>
* @param value The clusterId to set.
* @return This builder for chaining.
*/
public Builder setClusterId(
java.lang.String value) {
if (value == null) { throw new NullPointerException(); }
clusterId_ = value;
bitField0_ |= 0x00000002;
onChanged();
return this;
}
/**
* <code>optional string clusterId = 2;</code>
* @return This builder for chaining.
*/
public Builder clearClusterId() {
clusterId_ = getDefaultInstance().getClusterId();
bitField0_ = (bitField0_ & ~0x00000002);
onChanged();
return this;
}
/**
* <code>optional string clusterId = 2;</code>
* @param value The bytes for clusterId to set.
* @return This builder for chaining.
*/
public Builder setClusterIdBytes(
org.apache.hadoop.thirdparty.protobuf.ByteString value) {
if (value == null) { throw new NullPointerException(); }
clusterId_ = value;
bitField0_ |= 0x00000002;
onChanged();
return this;
}
private java.lang.Object nameserviceId_ = "";
/**
* <code>optional string nameserviceId = 3;</code>
* @return Whether the nameserviceId field is set.
*/
public boolean hasNameserviceId() {
return ((bitField0_ & 0x00000004) != 0);
}
/**
* <code>optional string nameserviceId = 3;</code>
* @return The nameserviceId.
*/
public java.lang.String getNameserviceId() {
java.lang.Object ref = nameserviceId_;
if (!(ref instanceof java.lang.String)) {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
nameserviceId_ = s;
}
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* <code>optional string nameserviceId = 3;</code>
* @return The bytes for nameserviceId.
*/
public org.apache.hadoop.thirdparty.protobuf.ByteString
getNameserviceIdBytes() {
java.lang.Object ref = nameserviceId_;
if (ref instanceof String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
nameserviceId_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
/**
* <code>optional string nameserviceId = 3;</code>
* @param value The nameserviceId to set.
* @return This builder for chaining.
*/
public Builder setNameserviceId(
java.lang.String value) {
if (value == null) { throw new NullPointerException(); }
nameserviceId_ = value;
bitField0_ |= 0x00000004;
onChanged();
return this;
}
/**
* <code>optional string nameserviceId = 3;</code>
* @return This builder for chaining.
*/
public Builder clearNameserviceId() {
nameserviceId_ = getDefaultInstance().getNameserviceId();
bitField0_ = (bitField0_ & ~0x00000004);
onChanged();
return this;
}
/**
* <code>optional string nameserviceId = 3;</code>
* @param value The bytes for nameserviceId to set.
* @return This builder for chaining.
*/
public Builder setNameserviceIdBytes(
org.apache.hadoop.thirdparty.protobuf.ByteString value) {
if (value == null) { throw new NullPointerException(); }
nameserviceId_ = value;
bitField0_ |= 0x00000004;
onChanged();
return this;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.FederationNamespaceInfoProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.FederationNamespaceInfoProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<FederationNamespaceInfoProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<FederationNamespaceInfoProto>() {
@java.lang.Override
public FederationNamespaceInfoProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<FederationNamespaceInfoProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<FederationNamespaceInfoProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface GetNamenodeRegistrationsRequestProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.GetNamenodeRegistrationsRequestProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* <code>optional .hadoop.hdfs.NamenodeMembershipRecordProto membership = 1;</code>
* @return Whether the membership field is set.
*/
boolean hasMembership();
/**
* <code>optional .hadoop.hdfs.NamenodeMembershipRecordProto membership = 1;</code>
* @return The membership.
*/
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto getMembership();
/**
* <code>optional .hadoop.hdfs.NamenodeMembershipRecordProto membership = 1;</code>
*/
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProtoOrBuilder getMembershipOrBuilder();
}
/**
* Protobuf type {@code hadoop.hdfs.GetNamenodeRegistrationsRequestProto}
*/
public static final class GetNamenodeRegistrationsRequestProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.GetNamenodeRegistrationsRequestProto)
GetNamenodeRegistrationsRequestProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use GetNamenodeRegistrationsRequestProto.newBuilder() to construct.
private GetNamenodeRegistrationsRequestProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private GetNamenodeRegistrationsRequestProto() {
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new GetNamenodeRegistrationsRequestProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetNamenodeRegistrationsRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetNamenodeRegistrationsRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsRequestProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsRequestProto.Builder.class);
}
private int bitField0_;
public static final int MEMBERSHIP_FIELD_NUMBER = 1;
private org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto membership_;
/**
* <code>optional .hadoop.hdfs.NamenodeMembershipRecordProto membership = 1;</code>
* @return Whether the membership field is set.
*/
@java.lang.Override
public boolean hasMembership() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional .hadoop.hdfs.NamenodeMembershipRecordProto membership = 1;</code>
* @return The membership.
*/
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto getMembership() {
return membership_ == null ? org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto.getDefaultInstance() : membership_;
}
/**
* <code>optional .hadoop.hdfs.NamenodeMembershipRecordProto membership = 1;</code>
*/
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProtoOrBuilder getMembershipOrBuilder() {
return membership_ == null ? org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto.getDefaultInstance() : membership_;
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) != 0)) {
output.writeMessage(1, getMembership());
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeMessageSize(1, getMembership());
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsRequestProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsRequestProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsRequestProto) obj;
if (hasMembership() != other.hasMembership()) return false;
if (hasMembership()) {
if (!getMembership()
.equals(other.getMembership())) return false;
}
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (hasMembership()) {
hash = (37 * hash) + MEMBERSHIP_FIELD_NUMBER;
hash = (53 * hash) + getMembership().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsRequestProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsRequestProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsRequestProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsRequestProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsRequestProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsRequestProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsRequestProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsRequestProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsRequestProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.GetNamenodeRegistrationsRequestProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.GetNamenodeRegistrationsRequestProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsRequestProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetNamenodeRegistrationsRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetNamenodeRegistrationsRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsRequestProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsRequestProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsRequestProto.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
if (org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
getMembershipFieldBuilder();
}
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
membership_ = null;
if (membershipBuilder_ != null) {
membershipBuilder_.dispose();
membershipBuilder_ = null;
}
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetNamenodeRegistrationsRequestProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsRequestProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsRequestProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsRequestProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsRequestProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsRequestProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsRequestProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsRequestProto(this);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartial0(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsRequestProto result) {
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) != 0)) {
result.membership_ = membershipBuilder_ == null
? membership_
: membershipBuilder_.build();
to_bitField0_ |= 0x00000001;
}
result.bitField0_ |= to_bitField0_;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsRequestProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsRequestProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsRequestProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsRequestProto.getDefaultInstance()) return this;
if (other.hasMembership()) {
mergeMembership(other.getMembership());
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 10: {
input.readMessage(
getMembershipFieldBuilder().getBuilder(),
extensionRegistry);
bitField0_ |= 0x00000001;
break;
} // case 10
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto membership_;
private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto.Builder, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProtoOrBuilder> membershipBuilder_;
/**
* <code>optional .hadoop.hdfs.NamenodeMembershipRecordProto membership = 1;</code>
* @return Whether the membership field is set.
*/
public boolean hasMembership() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional .hadoop.hdfs.NamenodeMembershipRecordProto membership = 1;</code>
* @return The membership.
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto getMembership() {
if (membershipBuilder_ == null) {
return membership_ == null ? org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto.getDefaultInstance() : membership_;
} else {
return membershipBuilder_.getMessage();
}
}
/**
* <code>optional .hadoop.hdfs.NamenodeMembershipRecordProto membership = 1;</code>
*/
public Builder setMembership(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto value) {
if (membershipBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
membership_ = value;
} else {
membershipBuilder_.setMessage(value);
}
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>optional .hadoop.hdfs.NamenodeMembershipRecordProto membership = 1;</code>
*/
public Builder setMembership(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto.Builder builderForValue) {
if (membershipBuilder_ == null) {
membership_ = builderForValue.build();
} else {
membershipBuilder_.setMessage(builderForValue.build());
}
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>optional .hadoop.hdfs.NamenodeMembershipRecordProto membership = 1;</code>
*/
public Builder mergeMembership(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto value) {
if (membershipBuilder_ == null) {
if (((bitField0_ & 0x00000001) != 0) &&
membership_ != null &&
membership_ != org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto.getDefaultInstance()) {
getMembershipBuilder().mergeFrom(value);
} else {
membership_ = value;
}
} else {
membershipBuilder_.mergeFrom(value);
}
if (membership_ != null) {
bitField0_ |= 0x00000001;
onChanged();
}
return this;
}
/**
* <code>optional .hadoop.hdfs.NamenodeMembershipRecordProto membership = 1;</code>
*/
public Builder clearMembership() {
bitField0_ = (bitField0_ & ~0x00000001);
membership_ = null;
if (membershipBuilder_ != null) {
membershipBuilder_.dispose();
membershipBuilder_ = null;
}
onChanged();
return this;
}
/**
* <code>optional .hadoop.hdfs.NamenodeMembershipRecordProto membership = 1;</code>
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto.Builder getMembershipBuilder() {
bitField0_ |= 0x00000001;
onChanged();
return getMembershipFieldBuilder().getBuilder();
}
/**
* <code>optional .hadoop.hdfs.NamenodeMembershipRecordProto membership = 1;</code>
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProtoOrBuilder getMembershipOrBuilder() {
if (membershipBuilder_ != null) {
return membershipBuilder_.getMessageOrBuilder();
} else {
return membership_ == null ?
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto.getDefaultInstance() : membership_;
}
}
/**
* <code>optional .hadoop.hdfs.NamenodeMembershipRecordProto membership = 1;</code>
*/
private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto.Builder, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProtoOrBuilder>
getMembershipFieldBuilder() {
if (membershipBuilder_ == null) {
membershipBuilder_ = new org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto.Builder, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProtoOrBuilder>(
getMembership(),
getParentForChildren(),
isClean());
membership_ = null;
}
return membershipBuilder_;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.GetNamenodeRegistrationsRequestProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.GetNamenodeRegistrationsRequestProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsRequestProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsRequestProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsRequestProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<GetNamenodeRegistrationsRequestProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<GetNamenodeRegistrationsRequestProto>() {
@java.lang.Override
public GetNamenodeRegistrationsRequestProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<GetNamenodeRegistrationsRequestProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<GetNamenodeRegistrationsRequestProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsRequestProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface GetNamenodeRegistrationsResponseProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.GetNamenodeRegistrationsResponseProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* <code>repeated .hadoop.hdfs.NamenodeMembershipRecordProto namenodeMemberships = 1;</code>
*/
java.util.List<org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto>
getNamenodeMembershipsList();
/**
* <code>repeated .hadoop.hdfs.NamenodeMembershipRecordProto namenodeMemberships = 1;</code>
*/
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto getNamenodeMemberships(int index);
/**
* <code>repeated .hadoop.hdfs.NamenodeMembershipRecordProto namenodeMemberships = 1;</code>
*/
int getNamenodeMembershipsCount();
/**
* <code>repeated .hadoop.hdfs.NamenodeMembershipRecordProto namenodeMemberships = 1;</code>
*/
java.util.List<? extends org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProtoOrBuilder>
getNamenodeMembershipsOrBuilderList();
/**
* <code>repeated .hadoop.hdfs.NamenodeMembershipRecordProto namenodeMemberships = 1;</code>
*/
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProtoOrBuilder getNamenodeMembershipsOrBuilder(
int index);
}
/**
* Protobuf type {@code hadoop.hdfs.GetNamenodeRegistrationsResponseProto}
*/
public static final class GetNamenodeRegistrationsResponseProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.GetNamenodeRegistrationsResponseProto)
GetNamenodeRegistrationsResponseProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use GetNamenodeRegistrationsResponseProto.newBuilder() to construct.
private GetNamenodeRegistrationsResponseProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private GetNamenodeRegistrationsResponseProto() {
namenodeMemberships_ = java.util.Collections.emptyList();
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new GetNamenodeRegistrationsResponseProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetNamenodeRegistrationsResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetNamenodeRegistrationsResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsResponseProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsResponseProto.Builder.class);
}
public static final int NAMENODEMEMBERSHIPS_FIELD_NUMBER = 1;
@SuppressWarnings("serial")
private java.util.List<org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto> namenodeMemberships_;
/**
* <code>repeated .hadoop.hdfs.NamenodeMembershipRecordProto namenodeMemberships = 1;</code>
*/
@java.lang.Override
public java.util.List<org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto> getNamenodeMembershipsList() {
return namenodeMemberships_;
}
/**
* <code>repeated .hadoop.hdfs.NamenodeMembershipRecordProto namenodeMemberships = 1;</code>
*/
@java.lang.Override
public java.util.List<? extends org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProtoOrBuilder>
getNamenodeMembershipsOrBuilderList() {
return namenodeMemberships_;
}
/**
* <code>repeated .hadoop.hdfs.NamenodeMembershipRecordProto namenodeMemberships = 1;</code>
*/
@java.lang.Override
public int getNamenodeMembershipsCount() {
return namenodeMemberships_.size();
}
/**
* <code>repeated .hadoop.hdfs.NamenodeMembershipRecordProto namenodeMemberships = 1;</code>
*/
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto getNamenodeMemberships(int index) {
return namenodeMemberships_.get(index);
}
/**
* <code>repeated .hadoop.hdfs.NamenodeMembershipRecordProto namenodeMemberships = 1;</code>
*/
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProtoOrBuilder getNamenodeMembershipsOrBuilder(
int index) {
return namenodeMemberships_.get(index);
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
for (int i = 0; i < namenodeMemberships_.size(); i++) {
output.writeMessage(1, namenodeMemberships_.get(i));
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
for (int i = 0; i < namenodeMemberships_.size(); i++) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeMessageSize(1, namenodeMemberships_.get(i));
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsResponseProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsResponseProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsResponseProto) obj;
if (!getNamenodeMembershipsList()
.equals(other.getNamenodeMembershipsList())) return false;
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (getNamenodeMembershipsCount() > 0) {
hash = (37 * hash) + NAMENODEMEMBERSHIPS_FIELD_NUMBER;
hash = (53 * hash) + getNamenodeMembershipsList().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsResponseProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsResponseProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsResponseProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsResponseProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsResponseProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsResponseProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsResponseProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsResponseProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsResponseProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.GetNamenodeRegistrationsResponseProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.GetNamenodeRegistrationsResponseProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsResponseProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetNamenodeRegistrationsResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetNamenodeRegistrationsResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsResponseProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsResponseProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsResponseProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
if (namenodeMembershipsBuilder_ == null) {
namenodeMemberships_ = java.util.Collections.emptyList();
} else {
namenodeMemberships_ = null;
namenodeMembershipsBuilder_.clear();
}
bitField0_ = (bitField0_ & ~0x00000001);
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetNamenodeRegistrationsResponseProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsResponseProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsResponseProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsResponseProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsResponseProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsResponseProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsResponseProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsResponseProto(this);
buildPartialRepeatedFields(result);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartialRepeatedFields(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsResponseProto result) {
if (namenodeMembershipsBuilder_ == null) {
if (((bitField0_ & 0x00000001) != 0)) {
namenodeMemberships_ = java.util.Collections.unmodifiableList(namenodeMemberships_);
bitField0_ = (bitField0_ & ~0x00000001);
}
result.namenodeMemberships_ = namenodeMemberships_;
} else {
result.namenodeMemberships_ = namenodeMembershipsBuilder_.build();
}
}
private void buildPartial0(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsResponseProto result) {
int from_bitField0_ = bitField0_;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsResponseProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsResponseProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsResponseProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsResponseProto.getDefaultInstance()) return this;
if (namenodeMembershipsBuilder_ == null) {
if (!other.namenodeMemberships_.isEmpty()) {
if (namenodeMemberships_.isEmpty()) {
namenodeMemberships_ = other.namenodeMemberships_;
bitField0_ = (bitField0_ & ~0x00000001);
} else {
ensureNamenodeMembershipsIsMutable();
namenodeMemberships_.addAll(other.namenodeMemberships_);
}
onChanged();
}
} else {
if (!other.namenodeMemberships_.isEmpty()) {
if (namenodeMembershipsBuilder_.isEmpty()) {
namenodeMembershipsBuilder_.dispose();
namenodeMembershipsBuilder_ = null;
namenodeMemberships_ = other.namenodeMemberships_;
bitField0_ = (bitField0_ & ~0x00000001);
namenodeMembershipsBuilder_ =
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
getNamenodeMembershipsFieldBuilder() : null;
} else {
namenodeMembershipsBuilder_.addAllMessages(other.namenodeMemberships_);
}
}
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 10: {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto m =
input.readMessage(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto.PARSER,
extensionRegistry);
if (namenodeMembershipsBuilder_ == null) {
ensureNamenodeMembershipsIsMutable();
namenodeMemberships_.add(m);
} else {
namenodeMembershipsBuilder_.addMessage(m);
}
break;
} // case 10
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private java.util.List<org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto> namenodeMemberships_ =
java.util.Collections.emptyList();
private void ensureNamenodeMembershipsIsMutable() {
if (!((bitField0_ & 0x00000001) != 0)) {
namenodeMemberships_ = new java.util.ArrayList<org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto>(namenodeMemberships_);
bitField0_ |= 0x00000001;
}
}
private org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto.Builder, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProtoOrBuilder> namenodeMembershipsBuilder_;
/**
* <code>repeated .hadoop.hdfs.NamenodeMembershipRecordProto namenodeMemberships = 1;</code>
*/
public java.util.List<org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto> getNamenodeMembershipsList() {
if (namenodeMembershipsBuilder_ == null) {
return java.util.Collections.unmodifiableList(namenodeMemberships_);
} else {
return namenodeMembershipsBuilder_.getMessageList();
}
}
/**
* <code>repeated .hadoop.hdfs.NamenodeMembershipRecordProto namenodeMemberships = 1;</code>
*/
public int getNamenodeMembershipsCount() {
if (namenodeMembershipsBuilder_ == null) {
return namenodeMemberships_.size();
} else {
return namenodeMembershipsBuilder_.getCount();
}
}
/**
* <code>repeated .hadoop.hdfs.NamenodeMembershipRecordProto namenodeMemberships = 1;</code>
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto getNamenodeMemberships(int index) {
if (namenodeMembershipsBuilder_ == null) {
return namenodeMemberships_.get(index);
} else {
return namenodeMembershipsBuilder_.getMessage(index);
}
}
/**
* <code>repeated .hadoop.hdfs.NamenodeMembershipRecordProto namenodeMemberships = 1;</code>
*/
public Builder setNamenodeMemberships(
int index, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto value) {
if (namenodeMembershipsBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
ensureNamenodeMembershipsIsMutable();
namenodeMemberships_.set(index, value);
onChanged();
} else {
namenodeMembershipsBuilder_.setMessage(index, value);
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.NamenodeMembershipRecordProto namenodeMemberships = 1;</code>
*/
public Builder setNamenodeMemberships(
int index, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto.Builder builderForValue) {
if (namenodeMembershipsBuilder_ == null) {
ensureNamenodeMembershipsIsMutable();
namenodeMemberships_.set(index, builderForValue.build());
onChanged();
} else {
namenodeMembershipsBuilder_.setMessage(index, builderForValue.build());
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.NamenodeMembershipRecordProto namenodeMemberships = 1;</code>
*/
public Builder addNamenodeMemberships(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto value) {
if (namenodeMembershipsBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
ensureNamenodeMembershipsIsMutable();
namenodeMemberships_.add(value);
onChanged();
} else {
namenodeMembershipsBuilder_.addMessage(value);
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.NamenodeMembershipRecordProto namenodeMemberships = 1;</code>
*/
public Builder addNamenodeMemberships(
int index, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto value) {
if (namenodeMembershipsBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
ensureNamenodeMembershipsIsMutable();
namenodeMemberships_.add(index, value);
onChanged();
} else {
namenodeMembershipsBuilder_.addMessage(index, value);
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.NamenodeMembershipRecordProto namenodeMemberships = 1;</code>
*/
public Builder addNamenodeMemberships(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto.Builder builderForValue) {
if (namenodeMembershipsBuilder_ == null) {
ensureNamenodeMembershipsIsMutable();
namenodeMemberships_.add(builderForValue.build());
onChanged();
} else {
namenodeMembershipsBuilder_.addMessage(builderForValue.build());
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.NamenodeMembershipRecordProto namenodeMemberships = 1;</code>
*/
public Builder addNamenodeMemberships(
int index, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto.Builder builderForValue) {
if (namenodeMembershipsBuilder_ == null) {
ensureNamenodeMembershipsIsMutable();
namenodeMemberships_.add(index, builderForValue.build());
onChanged();
} else {
namenodeMembershipsBuilder_.addMessage(index, builderForValue.build());
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.NamenodeMembershipRecordProto namenodeMemberships = 1;</code>
*/
public Builder addAllNamenodeMemberships(
java.lang.Iterable<? extends org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto> values) {
if (namenodeMembershipsBuilder_ == null) {
ensureNamenodeMembershipsIsMutable();
org.apache.hadoop.thirdparty.protobuf.AbstractMessageLite.Builder.addAll(
values, namenodeMemberships_);
onChanged();
} else {
namenodeMembershipsBuilder_.addAllMessages(values);
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.NamenodeMembershipRecordProto namenodeMemberships = 1;</code>
*/
public Builder clearNamenodeMemberships() {
if (namenodeMembershipsBuilder_ == null) {
namenodeMemberships_ = java.util.Collections.emptyList();
bitField0_ = (bitField0_ & ~0x00000001);
onChanged();
} else {
namenodeMembershipsBuilder_.clear();
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.NamenodeMembershipRecordProto namenodeMemberships = 1;</code>
*/
public Builder removeNamenodeMemberships(int index) {
if (namenodeMembershipsBuilder_ == null) {
ensureNamenodeMembershipsIsMutable();
namenodeMemberships_.remove(index);
onChanged();
} else {
namenodeMembershipsBuilder_.remove(index);
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.NamenodeMembershipRecordProto namenodeMemberships = 1;</code>
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto.Builder getNamenodeMembershipsBuilder(
int index) {
return getNamenodeMembershipsFieldBuilder().getBuilder(index);
}
/**
* <code>repeated .hadoop.hdfs.NamenodeMembershipRecordProto namenodeMemberships = 1;</code>
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProtoOrBuilder getNamenodeMembershipsOrBuilder(
int index) {
if (namenodeMembershipsBuilder_ == null) {
return namenodeMemberships_.get(index); } else {
return namenodeMembershipsBuilder_.getMessageOrBuilder(index);
}
}
/**
* <code>repeated .hadoop.hdfs.NamenodeMembershipRecordProto namenodeMemberships = 1;</code>
*/
public java.util.List<? extends org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProtoOrBuilder>
getNamenodeMembershipsOrBuilderList() {
if (namenodeMembershipsBuilder_ != null) {
return namenodeMembershipsBuilder_.getMessageOrBuilderList();
} else {
return java.util.Collections.unmodifiableList(namenodeMemberships_);
}
}
/**
* <code>repeated .hadoop.hdfs.NamenodeMembershipRecordProto namenodeMemberships = 1;</code>
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto.Builder addNamenodeMembershipsBuilder() {
return getNamenodeMembershipsFieldBuilder().addBuilder(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto.getDefaultInstance());
}
/**
* <code>repeated .hadoop.hdfs.NamenodeMembershipRecordProto namenodeMemberships = 1;</code>
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto.Builder addNamenodeMembershipsBuilder(
int index) {
return getNamenodeMembershipsFieldBuilder().addBuilder(
index, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto.getDefaultInstance());
}
/**
* <code>repeated .hadoop.hdfs.NamenodeMembershipRecordProto namenodeMemberships = 1;</code>
*/
public java.util.List<org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto.Builder>
getNamenodeMembershipsBuilderList() {
return getNamenodeMembershipsFieldBuilder().getBuilderList();
}
private org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto.Builder, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProtoOrBuilder>
getNamenodeMembershipsFieldBuilder() {
if (namenodeMembershipsBuilder_ == null) {
namenodeMembershipsBuilder_ = new org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto.Builder, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProtoOrBuilder>(
namenodeMemberships_,
((bitField0_ & 0x00000001) != 0),
getParentForChildren(),
isClean());
namenodeMemberships_ = null;
}
return namenodeMembershipsBuilder_;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.GetNamenodeRegistrationsResponseProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.GetNamenodeRegistrationsResponseProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsResponseProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsResponseProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsResponseProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<GetNamenodeRegistrationsResponseProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<GetNamenodeRegistrationsResponseProto>() {
@java.lang.Override
public GetNamenodeRegistrationsResponseProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<GetNamenodeRegistrationsResponseProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<GetNamenodeRegistrationsResponseProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamenodeRegistrationsResponseProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface GetExpiredRegistrationsRequestProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.GetExpiredRegistrationsRequestProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
}
/**
* Protobuf type {@code hadoop.hdfs.GetExpiredRegistrationsRequestProto}
*/
public static final class GetExpiredRegistrationsRequestProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.GetExpiredRegistrationsRequestProto)
GetExpiredRegistrationsRequestProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use GetExpiredRegistrationsRequestProto.newBuilder() to construct.
private GetExpiredRegistrationsRequestProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private GetExpiredRegistrationsRequestProto() {
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new GetExpiredRegistrationsRequestProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetExpiredRegistrationsRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetExpiredRegistrationsRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetExpiredRegistrationsRequestProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetExpiredRegistrationsRequestProto.Builder.class);
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetExpiredRegistrationsRequestProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetExpiredRegistrationsRequestProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetExpiredRegistrationsRequestProto) obj;
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetExpiredRegistrationsRequestProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetExpiredRegistrationsRequestProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetExpiredRegistrationsRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetExpiredRegistrationsRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetExpiredRegistrationsRequestProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetExpiredRegistrationsRequestProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetExpiredRegistrationsRequestProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetExpiredRegistrationsRequestProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetExpiredRegistrationsRequestProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetExpiredRegistrationsRequestProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetExpiredRegistrationsRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetExpiredRegistrationsRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetExpiredRegistrationsRequestProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.GetExpiredRegistrationsRequestProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.GetExpiredRegistrationsRequestProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetExpiredRegistrationsRequestProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetExpiredRegistrationsRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetExpiredRegistrationsRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetExpiredRegistrationsRequestProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetExpiredRegistrationsRequestProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetExpiredRegistrationsRequestProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetExpiredRegistrationsRequestProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetExpiredRegistrationsRequestProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetExpiredRegistrationsRequestProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetExpiredRegistrationsRequestProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetExpiredRegistrationsRequestProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetExpiredRegistrationsRequestProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetExpiredRegistrationsRequestProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetExpiredRegistrationsRequestProto(this);
onBuilt();
return result;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetExpiredRegistrationsRequestProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetExpiredRegistrationsRequestProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetExpiredRegistrationsRequestProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetExpiredRegistrationsRequestProto.getDefaultInstance()) return this;
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.GetExpiredRegistrationsRequestProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.GetExpiredRegistrationsRequestProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetExpiredRegistrationsRequestProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetExpiredRegistrationsRequestProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetExpiredRegistrationsRequestProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<GetExpiredRegistrationsRequestProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<GetExpiredRegistrationsRequestProto>() {
@java.lang.Override
public GetExpiredRegistrationsRequestProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<GetExpiredRegistrationsRequestProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<GetExpiredRegistrationsRequestProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetExpiredRegistrationsRequestProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface GetNamespaceInfoRequestProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.GetNamespaceInfoRequestProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
}
/**
* Protobuf type {@code hadoop.hdfs.GetNamespaceInfoRequestProto}
*/
public static final class GetNamespaceInfoRequestProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.GetNamespaceInfoRequestProto)
GetNamespaceInfoRequestProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use GetNamespaceInfoRequestProto.newBuilder() to construct.
private GetNamespaceInfoRequestProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private GetNamespaceInfoRequestProto() {
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new GetNamespaceInfoRequestProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetNamespaceInfoRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetNamespaceInfoRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoRequestProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoRequestProto.Builder.class);
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoRequestProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoRequestProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoRequestProto) obj;
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoRequestProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoRequestProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoRequestProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoRequestProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoRequestProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoRequestProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoRequestProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoRequestProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoRequestProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.GetNamespaceInfoRequestProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.GetNamespaceInfoRequestProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoRequestProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetNamespaceInfoRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetNamespaceInfoRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoRequestProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoRequestProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoRequestProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetNamespaceInfoRequestProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoRequestProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoRequestProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoRequestProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoRequestProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoRequestProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoRequestProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoRequestProto(this);
onBuilt();
return result;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoRequestProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoRequestProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoRequestProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoRequestProto.getDefaultInstance()) return this;
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.GetNamespaceInfoRequestProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.GetNamespaceInfoRequestProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoRequestProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoRequestProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoRequestProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<GetNamespaceInfoRequestProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<GetNamespaceInfoRequestProto>() {
@java.lang.Override
public GetNamespaceInfoRequestProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<GetNamespaceInfoRequestProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<GetNamespaceInfoRequestProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoRequestProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface GetNamespaceInfoResponseProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.GetNamespaceInfoResponseProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* <code>repeated .hadoop.hdfs.FederationNamespaceInfoProto namespaceInfos = 1;</code>
*/
java.util.List<org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto>
getNamespaceInfosList();
/**
* <code>repeated .hadoop.hdfs.FederationNamespaceInfoProto namespaceInfos = 1;</code>
*/
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto getNamespaceInfos(int index);
/**
* <code>repeated .hadoop.hdfs.FederationNamespaceInfoProto namespaceInfos = 1;</code>
*/
int getNamespaceInfosCount();
/**
* <code>repeated .hadoop.hdfs.FederationNamespaceInfoProto namespaceInfos = 1;</code>
*/
java.util.List<? extends org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProtoOrBuilder>
getNamespaceInfosOrBuilderList();
/**
* <code>repeated .hadoop.hdfs.FederationNamespaceInfoProto namespaceInfos = 1;</code>
*/
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProtoOrBuilder getNamespaceInfosOrBuilder(
int index);
}
/**
* Protobuf type {@code hadoop.hdfs.GetNamespaceInfoResponseProto}
*/
public static final class GetNamespaceInfoResponseProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.GetNamespaceInfoResponseProto)
GetNamespaceInfoResponseProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use GetNamespaceInfoResponseProto.newBuilder() to construct.
private GetNamespaceInfoResponseProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private GetNamespaceInfoResponseProto() {
namespaceInfos_ = java.util.Collections.emptyList();
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new GetNamespaceInfoResponseProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetNamespaceInfoResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetNamespaceInfoResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoResponseProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoResponseProto.Builder.class);
}
public static final int NAMESPACEINFOS_FIELD_NUMBER = 1;
@SuppressWarnings("serial")
private java.util.List<org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto> namespaceInfos_;
/**
* <code>repeated .hadoop.hdfs.FederationNamespaceInfoProto namespaceInfos = 1;</code>
*/
@java.lang.Override
public java.util.List<org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto> getNamespaceInfosList() {
return namespaceInfos_;
}
/**
* <code>repeated .hadoop.hdfs.FederationNamespaceInfoProto namespaceInfos = 1;</code>
*/
@java.lang.Override
public java.util.List<? extends org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProtoOrBuilder>
getNamespaceInfosOrBuilderList() {
return namespaceInfos_;
}
/**
* <code>repeated .hadoop.hdfs.FederationNamespaceInfoProto namespaceInfos = 1;</code>
*/
@java.lang.Override
public int getNamespaceInfosCount() {
return namespaceInfos_.size();
}
/**
* <code>repeated .hadoop.hdfs.FederationNamespaceInfoProto namespaceInfos = 1;</code>
*/
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto getNamespaceInfos(int index) {
return namespaceInfos_.get(index);
}
/**
* <code>repeated .hadoop.hdfs.FederationNamespaceInfoProto namespaceInfos = 1;</code>
*/
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProtoOrBuilder getNamespaceInfosOrBuilder(
int index) {
return namespaceInfos_.get(index);
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
for (int i = 0; i < namespaceInfos_.size(); i++) {
output.writeMessage(1, namespaceInfos_.get(i));
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
for (int i = 0; i < namespaceInfos_.size(); i++) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeMessageSize(1, namespaceInfos_.get(i));
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoResponseProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoResponseProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoResponseProto) obj;
if (!getNamespaceInfosList()
.equals(other.getNamespaceInfosList())) return false;
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (getNamespaceInfosCount() > 0) {
hash = (37 * hash) + NAMESPACEINFOS_FIELD_NUMBER;
hash = (53 * hash) + getNamespaceInfosList().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoResponseProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoResponseProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoResponseProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoResponseProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoResponseProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoResponseProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoResponseProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoResponseProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoResponseProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.GetNamespaceInfoResponseProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.GetNamespaceInfoResponseProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoResponseProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetNamespaceInfoResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetNamespaceInfoResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoResponseProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoResponseProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoResponseProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
if (namespaceInfosBuilder_ == null) {
namespaceInfos_ = java.util.Collections.emptyList();
} else {
namespaceInfos_ = null;
namespaceInfosBuilder_.clear();
}
bitField0_ = (bitField0_ & ~0x00000001);
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetNamespaceInfoResponseProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoResponseProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoResponseProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoResponseProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoResponseProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoResponseProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoResponseProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoResponseProto(this);
buildPartialRepeatedFields(result);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartialRepeatedFields(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoResponseProto result) {
if (namespaceInfosBuilder_ == null) {
if (((bitField0_ & 0x00000001) != 0)) {
namespaceInfos_ = java.util.Collections.unmodifiableList(namespaceInfos_);
bitField0_ = (bitField0_ & ~0x00000001);
}
result.namespaceInfos_ = namespaceInfos_;
} else {
result.namespaceInfos_ = namespaceInfosBuilder_.build();
}
}
private void buildPartial0(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoResponseProto result) {
int from_bitField0_ = bitField0_;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoResponseProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoResponseProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoResponseProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoResponseProto.getDefaultInstance()) return this;
if (namespaceInfosBuilder_ == null) {
if (!other.namespaceInfos_.isEmpty()) {
if (namespaceInfos_.isEmpty()) {
namespaceInfos_ = other.namespaceInfos_;
bitField0_ = (bitField0_ & ~0x00000001);
} else {
ensureNamespaceInfosIsMutable();
namespaceInfos_.addAll(other.namespaceInfos_);
}
onChanged();
}
} else {
if (!other.namespaceInfos_.isEmpty()) {
if (namespaceInfosBuilder_.isEmpty()) {
namespaceInfosBuilder_.dispose();
namespaceInfosBuilder_ = null;
namespaceInfos_ = other.namespaceInfos_;
bitField0_ = (bitField0_ & ~0x00000001);
namespaceInfosBuilder_ =
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
getNamespaceInfosFieldBuilder() : null;
} else {
namespaceInfosBuilder_.addAllMessages(other.namespaceInfos_);
}
}
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 10: {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto m =
input.readMessage(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto.PARSER,
extensionRegistry);
if (namespaceInfosBuilder_ == null) {
ensureNamespaceInfosIsMutable();
namespaceInfos_.add(m);
} else {
namespaceInfosBuilder_.addMessage(m);
}
break;
} // case 10
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private java.util.List<org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto> namespaceInfos_ =
java.util.Collections.emptyList();
private void ensureNamespaceInfosIsMutable() {
if (!((bitField0_ & 0x00000001) != 0)) {
namespaceInfos_ = new java.util.ArrayList<org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto>(namespaceInfos_);
bitField0_ |= 0x00000001;
}
}
private org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto.Builder, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProtoOrBuilder> namespaceInfosBuilder_;
/**
* <code>repeated .hadoop.hdfs.FederationNamespaceInfoProto namespaceInfos = 1;</code>
*/
public java.util.List<org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto> getNamespaceInfosList() {
if (namespaceInfosBuilder_ == null) {
return java.util.Collections.unmodifiableList(namespaceInfos_);
} else {
return namespaceInfosBuilder_.getMessageList();
}
}
/**
* <code>repeated .hadoop.hdfs.FederationNamespaceInfoProto namespaceInfos = 1;</code>
*/
public int getNamespaceInfosCount() {
if (namespaceInfosBuilder_ == null) {
return namespaceInfos_.size();
} else {
return namespaceInfosBuilder_.getCount();
}
}
/**
* <code>repeated .hadoop.hdfs.FederationNamespaceInfoProto namespaceInfos = 1;</code>
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto getNamespaceInfos(int index) {
if (namespaceInfosBuilder_ == null) {
return namespaceInfos_.get(index);
} else {
return namespaceInfosBuilder_.getMessage(index);
}
}
/**
* <code>repeated .hadoop.hdfs.FederationNamespaceInfoProto namespaceInfos = 1;</code>
*/
public Builder setNamespaceInfos(
int index, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto value) {
if (namespaceInfosBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
ensureNamespaceInfosIsMutable();
namespaceInfos_.set(index, value);
onChanged();
} else {
namespaceInfosBuilder_.setMessage(index, value);
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.FederationNamespaceInfoProto namespaceInfos = 1;</code>
*/
public Builder setNamespaceInfos(
int index, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto.Builder builderForValue) {
if (namespaceInfosBuilder_ == null) {
ensureNamespaceInfosIsMutable();
namespaceInfos_.set(index, builderForValue.build());
onChanged();
} else {
namespaceInfosBuilder_.setMessage(index, builderForValue.build());
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.FederationNamespaceInfoProto namespaceInfos = 1;</code>
*/
public Builder addNamespaceInfos(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto value) {
if (namespaceInfosBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
ensureNamespaceInfosIsMutable();
namespaceInfos_.add(value);
onChanged();
} else {
namespaceInfosBuilder_.addMessage(value);
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.FederationNamespaceInfoProto namespaceInfos = 1;</code>
*/
public Builder addNamespaceInfos(
int index, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto value) {
if (namespaceInfosBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
ensureNamespaceInfosIsMutable();
namespaceInfos_.add(index, value);
onChanged();
} else {
namespaceInfosBuilder_.addMessage(index, value);
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.FederationNamespaceInfoProto namespaceInfos = 1;</code>
*/
public Builder addNamespaceInfos(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto.Builder builderForValue) {
if (namespaceInfosBuilder_ == null) {
ensureNamespaceInfosIsMutable();
namespaceInfos_.add(builderForValue.build());
onChanged();
} else {
namespaceInfosBuilder_.addMessage(builderForValue.build());
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.FederationNamespaceInfoProto namespaceInfos = 1;</code>
*/
public Builder addNamespaceInfos(
int index, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto.Builder builderForValue) {
if (namespaceInfosBuilder_ == null) {
ensureNamespaceInfosIsMutable();
namespaceInfos_.add(index, builderForValue.build());
onChanged();
} else {
namespaceInfosBuilder_.addMessage(index, builderForValue.build());
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.FederationNamespaceInfoProto namespaceInfos = 1;</code>
*/
public Builder addAllNamespaceInfos(
java.lang.Iterable<? extends org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto> values) {
if (namespaceInfosBuilder_ == null) {
ensureNamespaceInfosIsMutable();
org.apache.hadoop.thirdparty.protobuf.AbstractMessageLite.Builder.addAll(
values, namespaceInfos_);
onChanged();
} else {
namespaceInfosBuilder_.addAllMessages(values);
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.FederationNamespaceInfoProto namespaceInfos = 1;</code>
*/
public Builder clearNamespaceInfos() {
if (namespaceInfosBuilder_ == null) {
namespaceInfos_ = java.util.Collections.emptyList();
bitField0_ = (bitField0_ & ~0x00000001);
onChanged();
} else {
namespaceInfosBuilder_.clear();
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.FederationNamespaceInfoProto namespaceInfos = 1;</code>
*/
public Builder removeNamespaceInfos(int index) {
if (namespaceInfosBuilder_ == null) {
ensureNamespaceInfosIsMutable();
namespaceInfos_.remove(index);
onChanged();
} else {
namespaceInfosBuilder_.remove(index);
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.FederationNamespaceInfoProto namespaceInfos = 1;</code>
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto.Builder getNamespaceInfosBuilder(
int index) {
return getNamespaceInfosFieldBuilder().getBuilder(index);
}
/**
* <code>repeated .hadoop.hdfs.FederationNamespaceInfoProto namespaceInfos = 1;</code>
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProtoOrBuilder getNamespaceInfosOrBuilder(
int index) {
if (namespaceInfosBuilder_ == null) {
return namespaceInfos_.get(index); } else {
return namespaceInfosBuilder_.getMessageOrBuilder(index);
}
}
/**
* <code>repeated .hadoop.hdfs.FederationNamespaceInfoProto namespaceInfos = 1;</code>
*/
public java.util.List<? extends org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProtoOrBuilder>
getNamespaceInfosOrBuilderList() {
if (namespaceInfosBuilder_ != null) {
return namespaceInfosBuilder_.getMessageOrBuilderList();
} else {
return java.util.Collections.unmodifiableList(namespaceInfos_);
}
}
/**
* <code>repeated .hadoop.hdfs.FederationNamespaceInfoProto namespaceInfos = 1;</code>
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto.Builder addNamespaceInfosBuilder() {
return getNamespaceInfosFieldBuilder().addBuilder(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto.getDefaultInstance());
}
/**
* <code>repeated .hadoop.hdfs.FederationNamespaceInfoProto namespaceInfos = 1;</code>
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto.Builder addNamespaceInfosBuilder(
int index) {
return getNamespaceInfosFieldBuilder().addBuilder(
index, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto.getDefaultInstance());
}
/**
* <code>repeated .hadoop.hdfs.FederationNamespaceInfoProto namespaceInfos = 1;</code>
*/
public java.util.List<org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto.Builder>
getNamespaceInfosBuilderList() {
return getNamespaceInfosFieldBuilder().getBuilderList();
}
private org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto.Builder, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProtoOrBuilder>
getNamespaceInfosFieldBuilder() {
if (namespaceInfosBuilder_ == null) {
namespaceInfosBuilder_ = new org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProto.Builder, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.FederationNamespaceInfoProtoOrBuilder>(
namespaceInfos_,
((bitField0_ & 0x00000001) != 0),
getParentForChildren(),
isClean());
namespaceInfos_ = null;
}
return namespaceInfosBuilder_;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.GetNamespaceInfoResponseProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.GetNamespaceInfoResponseProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoResponseProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoResponseProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoResponseProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<GetNamespaceInfoResponseProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<GetNamespaceInfoResponseProto>() {
@java.lang.Override
public GetNamespaceInfoResponseProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<GetNamespaceInfoResponseProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<GetNamespaceInfoResponseProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetNamespaceInfoResponseProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface UpdateNamenodeRegistrationRequestProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.UpdateNamenodeRegistrationRequestProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* <code>optional string nameserviceId = 1;</code>
* @return Whether the nameserviceId field is set.
*/
boolean hasNameserviceId();
/**
* <code>optional string nameserviceId = 1;</code>
* @return The nameserviceId.
*/
java.lang.String getNameserviceId();
/**
* <code>optional string nameserviceId = 1;</code>
* @return The bytes for nameserviceId.
*/
org.apache.hadoop.thirdparty.protobuf.ByteString
getNameserviceIdBytes();
/**
* <code>optional string namenodeId = 2;</code>
* @return Whether the namenodeId field is set.
*/
boolean hasNamenodeId();
/**
* <code>optional string namenodeId = 2;</code>
* @return The namenodeId.
*/
java.lang.String getNamenodeId();
/**
* <code>optional string namenodeId = 2;</code>
* @return The bytes for namenodeId.
*/
org.apache.hadoop.thirdparty.protobuf.ByteString
getNamenodeIdBytes();
/**
* <code>optional string state = 3;</code>
* @return Whether the state field is set.
*/
boolean hasState();
/**
* <code>optional string state = 3;</code>
* @return The state.
*/
java.lang.String getState();
/**
* <code>optional string state = 3;</code>
* @return The bytes for state.
*/
org.apache.hadoop.thirdparty.protobuf.ByteString
getStateBytes();
}
/**
* Protobuf type {@code hadoop.hdfs.UpdateNamenodeRegistrationRequestProto}
*/
public static final class UpdateNamenodeRegistrationRequestProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.UpdateNamenodeRegistrationRequestProto)
UpdateNamenodeRegistrationRequestProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use UpdateNamenodeRegistrationRequestProto.newBuilder() to construct.
private UpdateNamenodeRegistrationRequestProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private UpdateNamenodeRegistrationRequestProto() {
nameserviceId_ = "";
namenodeId_ = "";
state_ = "";
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new UpdateNamenodeRegistrationRequestProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_UpdateNamenodeRegistrationRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_UpdateNamenodeRegistrationRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationRequestProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationRequestProto.Builder.class);
}
private int bitField0_;
public static final int NAMESERVICEID_FIELD_NUMBER = 1;
@SuppressWarnings("serial")
private volatile java.lang.Object nameserviceId_ = "";
/**
* <code>optional string nameserviceId = 1;</code>
* @return Whether the nameserviceId field is set.
*/
@java.lang.Override
public boolean hasNameserviceId() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional string nameserviceId = 1;</code>
* @return The nameserviceId.
*/
@java.lang.Override
public java.lang.String getNameserviceId() {
java.lang.Object ref = nameserviceId_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
nameserviceId_ = s;
}
return s;
}
}
/**
* <code>optional string nameserviceId = 1;</code>
* @return The bytes for nameserviceId.
*/
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.ByteString
getNameserviceIdBytes() {
java.lang.Object ref = nameserviceId_;
if (ref instanceof java.lang.String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
nameserviceId_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
public static final int NAMENODEID_FIELD_NUMBER = 2;
@SuppressWarnings("serial")
private volatile java.lang.Object namenodeId_ = "";
/**
* <code>optional string namenodeId = 2;</code>
* @return Whether the namenodeId field is set.
*/
@java.lang.Override
public boolean hasNamenodeId() {
return ((bitField0_ & 0x00000002) != 0);
}
/**
* <code>optional string namenodeId = 2;</code>
* @return The namenodeId.
*/
@java.lang.Override
public java.lang.String getNamenodeId() {
java.lang.Object ref = namenodeId_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
namenodeId_ = s;
}
return s;
}
}
/**
* <code>optional string namenodeId = 2;</code>
* @return The bytes for namenodeId.
*/
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.ByteString
getNamenodeIdBytes() {
java.lang.Object ref = namenodeId_;
if (ref instanceof java.lang.String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
namenodeId_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
public static final int STATE_FIELD_NUMBER = 3;
@SuppressWarnings("serial")
private volatile java.lang.Object state_ = "";
/**
* <code>optional string state = 3;</code>
* @return Whether the state field is set.
*/
@java.lang.Override
public boolean hasState() {
return ((bitField0_ & 0x00000004) != 0);
}
/**
* <code>optional string state = 3;</code>
* @return The state.
*/
@java.lang.Override
public java.lang.String getState() {
java.lang.Object ref = state_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
state_ = s;
}
return s;
}
}
/**
* <code>optional string state = 3;</code>
* @return The bytes for state.
*/
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.ByteString
getStateBytes() {
java.lang.Object ref = state_;
if (ref instanceof java.lang.String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
state_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) != 0)) {
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 1, nameserviceId_);
}
if (((bitField0_ & 0x00000002) != 0)) {
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 2, namenodeId_);
}
if (((bitField0_ & 0x00000004) != 0)) {
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 3, state_);
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(1, nameserviceId_);
}
if (((bitField0_ & 0x00000002) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(2, namenodeId_);
}
if (((bitField0_ & 0x00000004) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(3, state_);
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationRequestProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationRequestProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationRequestProto) obj;
if (hasNameserviceId() != other.hasNameserviceId()) return false;
if (hasNameserviceId()) {
if (!getNameserviceId()
.equals(other.getNameserviceId())) return false;
}
if (hasNamenodeId() != other.hasNamenodeId()) return false;
if (hasNamenodeId()) {
if (!getNamenodeId()
.equals(other.getNamenodeId())) return false;
}
if (hasState() != other.hasState()) return false;
if (hasState()) {
if (!getState()
.equals(other.getState())) return false;
}
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (hasNameserviceId()) {
hash = (37 * hash) + NAMESERVICEID_FIELD_NUMBER;
hash = (53 * hash) + getNameserviceId().hashCode();
}
if (hasNamenodeId()) {
hash = (37 * hash) + NAMENODEID_FIELD_NUMBER;
hash = (53 * hash) + getNamenodeId().hashCode();
}
if (hasState()) {
hash = (37 * hash) + STATE_FIELD_NUMBER;
hash = (53 * hash) + getState().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationRequestProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationRequestProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationRequestProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationRequestProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationRequestProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationRequestProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationRequestProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationRequestProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationRequestProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.UpdateNamenodeRegistrationRequestProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.UpdateNamenodeRegistrationRequestProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationRequestProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_UpdateNamenodeRegistrationRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_UpdateNamenodeRegistrationRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationRequestProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationRequestProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationRequestProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
nameserviceId_ = "";
namenodeId_ = "";
state_ = "";
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_UpdateNamenodeRegistrationRequestProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationRequestProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationRequestProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationRequestProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationRequestProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationRequestProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationRequestProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationRequestProto(this);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartial0(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationRequestProto result) {
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) != 0)) {
result.nameserviceId_ = nameserviceId_;
to_bitField0_ |= 0x00000001;
}
if (((from_bitField0_ & 0x00000002) != 0)) {
result.namenodeId_ = namenodeId_;
to_bitField0_ |= 0x00000002;
}
if (((from_bitField0_ & 0x00000004) != 0)) {
result.state_ = state_;
to_bitField0_ |= 0x00000004;
}
result.bitField0_ |= to_bitField0_;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationRequestProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationRequestProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationRequestProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationRequestProto.getDefaultInstance()) return this;
if (other.hasNameserviceId()) {
nameserviceId_ = other.nameserviceId_;
bitField0_ |= 0x00000001;
onChanged();
}
if (other.hasNamenodeId()) {
namenodeId_ = other.namenodeId_;
bitField0_ |= 0x00000002;
onChanged();
}
if (other.hasState()) {
state_ = other.state_;
bitField0_ |= 0x00000004;
onChanged();
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 10: {
nameserviceId_ = input.readBytes();
bitField0_ |= 0x00000001;
break;
} // case 10
case 18: {
namenodeId_ = input.readBytes();
bitField0_ |= 0x00000002;
break;
} // case 18
case 26: {
state_ = input.readBytes();
bitField0_ |= 0x00000004;
break;
} // case 26
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private java.lang.Object nameserviceId_ = "";
/**
* <code>optional string nameserviceId = 1;</code>
* @return Whether the nameserviceId field is set.
*/
public boolean hasNameserviceId() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional string nameserviceId = 1;</code>
* @return The nameserviceId.
*/
public java.lang.String getNameserviceId() {
java.lang.Object ref = nameserviceId_;
if (!(ref instanceof java.lang.String)) {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
nameserviceId_ = s;
}
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* <code>optional string nameserviceId = 1;</code>
* @return The bytes for nameserviceId.
*/
public org.apache.hadoop.thirdparty.protobuf.ByteString
getNameserviceIdBytes() {
java.lang.Object ref = nameserviceId_;
if (ref instanceof String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
nameserviceId_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
/**
* <code>optional string nameserviceId = 1;</code>
* @param value The nameserviceId to set.
* @return This builder for chaining.
*/
public Builder setNameserviceId(
java.lang.String value) {
if (value == null) { throw new NullPointerException(); }
nameserviceId_ = value;
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>optional string nameserviceId = 1;</code>
* @return This builder for chaining.
*/
public Builder clearNameserviceId() {
nameserviceId_ = getDefaultInstance().getNameserviceId();
bitField0_ = (bitField0_ & ~0x00000001);
onChanged();
return this;
}
/**
* <code>optional string nameserviceId = 1;</code>
* @param value The bytes for nameserviceId to set.
* @return This builder for chaining.
*/
public Builder setNameserviceIdBytes(
org.apache.hadoop.thirdparty.protobuf.ByteString value) {
if (value == null) { throw new NullPointerException(); }
nameserviceId_ = value;
bitField0_ |= 0x00000001;
onChanged();
return this;
}
private java.lang.Object namenodeId_ = "";
/**
* <code>optional string namenodeId = 2;</code>
* @return Whether the namenodeId field is set.
*/
public boolean hasNamenodeId() {
return ((bitField0_ & 0x00000002) != 0);
}
/**
* <code>optional string namenodeId = 2;</code>
* @return The namenodeId.
*/
public java.lang.String getNamenodeId() {
java.lang.Object ref = namenodeId_;
if (!(ref instanceof java.lang.String)) {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
namenodeId_ = s;
}
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* <code>optional string namenodeId = 2;</code>
* @return The bytes for namenodeId.
*/
public org.apache.hadoop.thirdparty.protobuf.ByteString
getNamenodeIdBytes() {
java.lang.Object ref = namenodeId_;
if (ref instanceof String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
namenodeId_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
/**
* <code>optional string namenodeId = 2;</code>
* @param value The namenodeId to set.
* @return This builder for chaining.
*/
public Builder setNamenodeId(
java.lang.String value) {
if (value == null) { throw new NullPointerException(); }
namenodeId_ = value;
bitField0_ |= 0x00000002;
onChanged();
return this;
}
/**
* <code>optional string namenodeId = 2;</code>
* @return This builder for chaining.
*/
public Builder clearNamenodeId() {
namenodeId_ = getDefaultInstance().getNamenodeId();
bitField0_ = (bitField0_ & ~0x00000002);
onChanged();
return this;
}
/**
* <code>optional string namenodeId = 2;</code>
* @param value The bytes for namenodeId to set.
* @return This builder for chaining.
*/
public Builder setNamenodeIdBytes(
org.apache.hadoop.thirdparty.protobuf.ByteString value) {
if (value == null) { throw new NullPointerException(); }
namenodeId_ = value;
bitField0_ |= 0x00000002;
onChanged();
return this;
}
private java.lang.Object state_ = "";
/**
* <code>optional string state = 3;</code>
* @return Whether the state field is set.
*/
public boolean hasState() {
return ((bitField0_ & 0x00000004) != 0);
}
/**
* <code>optional string state = 3;</code>
* @return The state.
*/
public java.lang.String getState() {
java.lang.Object ref = state_;
if (!(ref instanceof java.lang.String)) {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
state_ = s;
}
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* <code>optional string state = 3;</code>
* @return The bytes for state.
*/
public org.apache.hadoop.thirdparty.protobuf.ByteString
getStateBytes() {
java.lang.Object ref = state_;
if (ref instanceof String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
state_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
/**
* <code>optional string state = 3;</code>
* @param value The state to set.
* @return This builder for chaining.
*/
public Builder setState(
java.lang.String value) {
if (value == null) { throw new NullPointerException(); }
state_ = value;
bitField0_ |= 0x00000004;
onChanged();
return this;
}
/**
* <code>optional string state = 3;</code>
* @return This builder for chaining.
*/
public Builder clearState() {
state_ = getDefaultInstance().getState();
bitField0_ = (bitField0_ & ~0x00000004);
onChanged();
return this;
}
/**
* <code>optional string state = 3;</code>
* @param value The bytes for state to set.
* @return This builder for chaining.
*/
public Builder setStateBytes(
org.apache.hadoop.thirdparty.protobuf.ByteString value) {
if (value == null) { throw new NullPointerException(); }
state_ = value;
bitField0_ |= 0x00000004;
onChanged();
return this;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.UpdateNamenodeRegistrationRequestProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.UpdateNamenodeRegistrationRequestProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationRequestProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationRequestProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationRequestProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<UpdateNamenodeRegistrationRequestProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<UpdateNamenodeRegistrationRequestProto>() {
@java.lang.Override
public UpdateNamenodeRegistrationRequestProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<UpdateNamenodeRegistrationRequestProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<UpdateNamenodeRegistrationRequestProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationRequestProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface UpdateNamenodeRegistrationResponseProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.UpdateNamenodeRegistrationResponseProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* <code>optional bool status = 1;</code>
* @return Whether the status field is set.
*/
boolean hasStatus();
/**
* <code>optional bool status = 1;</code>
* @return The status.
*/
boolean getStatus();
}
/**
* Protobuf type {@code hadoop.hdfs.UpdateNamenodeRegistrationResponseProto}
*/
public static final class UpdateNamenodeRegistrationResponseProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.UpdateNamenodeRegistrationResponseProto)
UpdateNamenodeRegistrationResponseProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use UpdateNamenodeRegistrationResponseProto.newBuilder() to construct.
private UpdateNamenodeRegistrationResponseProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private UpdateNamenodeRegistrationResponseProto() {
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new UpdateNamenodeRegistrationResponseProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_UpdateNamenodeRegistrationResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_UpdateNamenodeRegistrationResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationResponseProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationResponseProto.Builder.class);
}
private int bitField0_;
public static final int STATUS_FIELD_NUMBER = 1;
private boolean status_ = false;
/**
* <code>optional bool status = 1;</code>
* @return Whether the status field is set.
*/
@java.lang.Override
public boolean hasStatus() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional bool status = 1;</code>
* @return The status.
*/
@java.lang.Override
public boolean getStatus() {
return status_;
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) != 0)) {
output.writeBool(1, status_);
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeBoolSize(1, status_);
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationResponseProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationResponseProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationResponseProto) obj;
if (hasStatus() != other.hasStatus()) return false;
if (hasStatus()) {
if (getStatus()
!= other.getStatus()) return false;
}
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (hasStatus()) {
hash = (37 * hash) + STATUS_FIELD_NUMBER;
hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashBoolean(
getStatus());
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationResponseProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationResponseProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationResponseProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationResponseProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationResponseProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationResponseProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationResponseProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationResponseProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationResponseProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.UpdateNamenodeRegistrationResponseProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.UpdateNamenodeRegistrationResponseProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationResponseProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_UpdateNamenodeRegistrationResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_UpdateNamenodeRegistrationResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationResponseProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationResponseProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationResponseProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
status_ = false;
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_UpdateNamenodeRegistrationResponseProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationResponseProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationResponseProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationResponseProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationResponseProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationResponseProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationResponseProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationResponseProto(this);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartial0(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationResponseProto result) {
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) != 0)) {
result.status_ = status_;
to_bitField0_ |= 0x00000001;
}
result.bitField0_ |= to_bitField0_;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationResponseProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationResponseProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationResponseProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationResponseProto.getDefaultInstance()) return this;
if (other.hasStatus()) {
setStatus(other.getStatus());
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 8: {
status_ = input.readBool();
bitField0_ |= 0x00000001;
break;
} // case 8
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private boolean status_ ;
/**
* <code>optional bool status = 1;</code>
* @return Whether the status field is set.
*/
@java.lang.Override
public boolean hasStatus() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional bool status = 1;</code>
* @return The status.
*/
@java.lang.Override
public boolean getStatus() {
return status_;
}
/**
* <code>optional bool status = 1;</code>
* @param value The status to set.
* @return This builder for chaining.
*/
public Builder setStatus(boolean value) {
status_ = value;
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>optional bool status = 1;</code>
* @return This builder for chaining.
*/
public Builder clearStatus() {
bitField0_ = (bitField0_ & ~0x00000001);
status_ = false;
onChanged();
return this;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.UpdateNamenodeRegistrationResponseProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.UpdateNamenodeRegistrationResponseProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationResponseProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationResponseProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationResponseProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<UpdateNamenodeRegistrationResponseProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<UpdateNamenodeRegistrationResponseProto>() {
@java.lang.Override
public UpdateNamenodeRegistrationResponseProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<UpdateNamenodeRegistrationResponseProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<UpdateNamenodeRegistrationResponseProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateNamenodeRegistrationResponseProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface NamenodeHeartbeatRequestProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.NamenodeHeartbeatRequestProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* <code>optional .hadoop.hdfs.NamenodeMembershipRecordProto namenodeMembership = 1;</code>
* @return Whether the namenodeMembership field is set.
*/
boolean hasNamenodeMembership();
/**
* <code>optional .hadoop.hdfs.NamenodeMembershipRecordProto namenodeMembership = 1;</code>
* @return The namenodeMembership.
*/
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto getNamenodeMembership();
/**
* <code>optional .hadoop.hdfs.NamenodeMembershipRecordProto namenodeMembership = 1;</code>
*/
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProtoOrBuilder getNamenodeMembershipOrBuilder();
}
/**
* Protobuf type {@code hadoop.hdfs.NamenodeHeartbeatRequestProto}
*/
public static final class NamenodeHeartbeatRequestProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.NamenodeHeartbeatRequestProto)
NamenodeHeartbeatRequestProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use NamenodeHeartbeatRequestProto.newBuilder() to construct.
private NamenodeHeartbeatRequestProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private NamenodeHeartbeatRequestProto() {
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new NamenodeHeartbeatRequestProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_NamenodeHeartbeatRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_NamenodeHeartbeatRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatRequestProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatRequestProto.Builder.class);
}
private int bitField0_;
public static final int NAMENODEMEMBERSHIP_FIELD_NUMBER = 1;
private org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto namenodeMembership_;
/**
* <code>optional .hadoop.hdfs.NamenodeMembershipRecordProto namenodeMembership = 1;</code>
* @return Whether the namenodeMembership field is set.
*/
@java.lang.Override
public boolean hasNamenodeMembership() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional .hadoop.hdfs.NamenodeMembershipRecordProto namenodeMembership = 1;</code>
* @return The namenodeMembership.
*/
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto getNamenodeMembership() {
return namenodeMembership_ == null ? org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto.getDefaultInstance() : namenodeMembership_;
}
/**
* <code>optional .hadoop.hdfs.NamenodeMembershipRecordProto namenodeMembership = 1;</code>
*/
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProtoOrBuilder getNamenodeMembershipOrBuilder() {
return namenodeMembership_ == null ? org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto.getDefaultInstance() : namenodeMembership_;
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) != 0)) {
output.writeMessage(1, getNamenodeMembership());
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeMessageSize(1, getNamenodeMembership());
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatRequestProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatRequestProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatRequestProto) obj;
if (hasNamenodeMembership() != other.hasNamenodeMembership()) return false;
if (hasNamenodeMembership()) {
if (!getNamenodeMembership()
.equals(other.getNamenodeMembership())) return false;
}
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (hasNamenodeMembership()) {
hash = (37 * hash) + NAMENODEMEMBERSHIP_FIELD_NUMBER;
hash = (53 * hash) + getNamenodeMembership().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatRequestProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatRequestProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatRequestProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatRequestProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatRequestProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatRequestProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatRequestProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatRequestProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatRequestProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.NamenodeHeartbeatRequestProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.NamenodeHeartbeatRequestProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatRequestProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_NamenodeHeartbeatRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_NamenodeHeartbeatRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatRequestProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatRequestProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatRequestProto.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
if (org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
getNamenodeMembershipFieldBuilder();
}
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
namenodeMembership_ = null;
if (namenodeMembershipBuilder_ != null) {
namenodeMembershipBuilder_.dispose();
namenodeMembershipBuilder_ = null;
}
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_NamenodeHeartbeatRequestProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatRequestProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatRequestProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatRequestProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatRequestProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatRequestProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatRequestProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatRequestProto(this);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartial0(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatRequestProto result) {
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) != 0)) {
result.namenodeMembership_ = namenodeMembershipBuilder_ == null
? namenodeMembership_
: namenodeMembershipBuilder_.build();
to_bitField0_ |= 0x00000001;
}
result.bitField0_ |= to_bitField0_;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatRequestProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatRequestProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatRequestProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatRequestProto.getDefaultInstance()) return this;
if (other.hasNamenodeMembership()) {
mergeNamenodeMembership(other.getNamenodeMembership());
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 10: {
input.readMessage(
getNamenodeMembershipFieldBuilder().getBuilder(),
extensionRegistry);
bitField0_ |= 0x00000001;
break;
} // case 10
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto namenodeMembership_;
private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto.Builder, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProtoOrBuilder> namenodeMembershipBuilder_;
/**
* <code>optional .hadoop.hdfs.NamenodeMembershipRecordProto namenodeMembership = 1;</code>
* @return Whether the namenodeMembership field is set.
*/
public boolean hasNamenodeMembership() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional .hadoop.hdfs.NamenodeMembershipRecordProto namenodeMembership = 1;</code>
* @return The namenodeMembership.
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto getNamenodeMembership() {
if (namenodeMembershipBuilder_ == null) {
return namenodeMembership_ == null ? org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto.getDefaultInstance() : namenodeMembership_;
} else {
return namenodeMembershipBuilder_.getMessage();
}
}
/**
* <code>optional .hadoop.hdfs.NamenodeMembershipRecordProto namenodeMembership = 1;</code>
*/
public Builder setNamenodeMembership(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto value) {
if (namenodeMembershipBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
namenodeMembership_ = value;
} else {
namenodeMembershipBuilder_.setMessage(value);
}
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>optional .hadoop.hdfs.NamenodeMembershipRecordProto namenodeMembership = 1;</code>
*/
public Builder setNamenodeMembership(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto.Builder builderForValue) {
if (namenodeMembershipBuilder_ == null) {
namenodeMembership_ = builderForValue.build();
} else {
namenodeMembershipBuilder_.setMessage(builderForValue.build());
}
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>optional .hadoop.hdfs.NamenodeMembershipRecordProto namenodeMembership = 1;</code>
*/
public Builder mergeNamenodeMembership(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto value) {
if (namenodeMembershipBuilder_ == null) {
if (((bitField0_ & 0x00000001) != 0) &&
namenodeMembership_ != null &&
namenodeMembership_ != org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto.getDefaultInstance()) {
getNamenodeMembershipBuilder().mergeFrom(value);
} else {
namenodeMembership_ = value;
}
} else {
namenodeMembershipBuilder_.mergeFrom(value);
}
if (namenodeMembership_ != null) {
bitField0_ |= 0x00000001;
onChanged();
}
return this;
}
/**
* <code>optional .hadoop.hdfs.NamenodeMembershipRecordProto namenodeMembership = 1;</code>
*/
public Builder clearNamenodeMembership() {
bitField0_ = (bitField0_ & ~0x00000001);
namenodeMembership_ = null;
if (namenodeMembershipBuilder_ != null) {
namenodeMembershipBuilder_.dispose();
namenodeMembershipBuilder_ = null;
}
onChanged();
return this;
}
/**
* <code>optional .hadoop.hdfs.NamenodeMembershipRecordProto namenodeMembership = 1;</code>
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto.Builder getNamenodeMembershipBuilder() {
bitField0_ |= 0x00000001;
onChanged();
return getNamenodeMembershipFieldBuilder().getBuilder();
}
/**
* <code>optional .hadoop.hdfs.NamenodeMembershipRecordProto namenodeMembership = 1;</code>
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProtoOrBuilder getNamenodeMembershipOrBuilder() {
if (namenodeMembershipBuilder_ != null) {
return namenodeMembershipBuilder_.getMessageOrBuilder();
} else {
return namenodeMembership_ == null ?
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto.getDefaultInstance() : namenodeMembership_;
}
}
/**
* <code>optional .hadoop.hdfs.NamenodeMembershipRecordProto namenodeMembership = 1;</code>
*/
private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto.Builder, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProtoOrBuilder>
getNamenodeMembershipFieldBuilder() {
if (namenodeMembershipBuilder_ == null) {
namenodeMembershipBuilder_ = new org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProto.Builder, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeMembershipRecordProtoOrBuilder>(
getNamenodeMembership(),
getParentForChildren(),
isClean());
namenodeMembership_ = null;
}
return namenodeMembershipBuilder_;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.NamenodeHeartbeatRequestProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.NamenodeHeartbeatRequestProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatRequestProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatRequestProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatRequestProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<NamenodeHeartbeatRequestProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<NamenodeHeartbeatRequestProto>() {
@java.lang.Override
public NamenodeHeartbeatRequestProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<NamenodeHeartbeatRequestProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<NamenodeHeartbeatRequestProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatRequestProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface NamenodeHeartbeatResponseProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.NamenodeHeartbeatResponseProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* <code>optional bool status = 1;</code>
* @return Whether the status field is set.
*/
boolean hasStatus();
/**
* <code>optional bool status = 1;</code>
* @return The status.
*/
boolean getStatus();
}
/**
* Protobuf type {@code hadoop.hdfs.NamenodeHeartbeatResponseProto}
*/
public static final class NamenodeHeartbeatResponseProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.NamenodeHeartbeatResponseProto)
NamenodeHeartbeatResponseProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use NamenodeHeartbeatResponseProto.newBuilder() to construct.
private NamenodeHeartbeatResponseProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private NamenodeHeartbeatResponseProto() {
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new NamenodeHeartbeatResponseProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_NamenodeHeartbeatResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_NamenodeHeartbeatResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatResponseProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatResponseProto.Builder.class);
}
private int bitField0_;
public static final int STATUS_FIELD_NUMBER = 1;
private boolean status_ = false;
/**
* <code>optional bool status = 1;</code>
* @return Whether the status field is set.
*/
@java.lang.Override
public boolean hasStatus() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional bool status = 1;</code>
* @return The status.
*/
@java.lang.Override
public boolean getStatus() {
return status_;
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) != 0)) {
output.writeBool(1, status_);
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeBoolSize(1, status_);
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatResponseProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatResponseProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatResponseProto) obj;
if (hasStatus() != other.hasStatus()) return false;
if (hasStatus()) {
if (getStatus()
!= other.getStatus()) return false;
}
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (hasStatus()) {
hash = (37 * hash) + STATUS_FIELD_NUMBER;
hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashBoolean(
getStatus());
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatResponseProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatResponseProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatResponseProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatResponseProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatResponseProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatResponseProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatResponseProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatResponseProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatResponseProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.NamenodeHeartbeatResponseProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.NamenodeHeartbeatResponseProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatResponseProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_NamenodeHeartbeatResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_NamenodeHeartbeatResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatResponseProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatResponseProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatResponseProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
status_ = false;
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_NamenodeHeartbeatResponseProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatResponseProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatResponseProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatResponseProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatResponseProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatResponseProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatResponseProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatResponseProto(this);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartial0(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatResponseProto result) {
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) != 0)) {
result.status_ = status_;
to_bitField0_ |= 0x00000001;
}
result.bitField0_ |= to_bitField0_;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatResponseProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatResponseProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatResponseProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatResponseProto.getDefaultInstance()) return this;
if (other.hasStatus()) {
setStatus(other.getStatus());
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 8: {
status_ = input.readBool();
bitField0_ |= 0x00000001;
break;
} // case 8
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private boolean status_ ;
/**
* <code>optional bool status = 1;</code>
* @return Whether the status field is set.
*/
@java.lang.Override
public boolean hasStatus() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional bool status = 1;</code>
* @return The status.
*/
@java.lang.Override
public boolean getStatus() {
return status_;
}
/**
* <code>optional bool status = 1;</code>
* @param value The status to set.
* @return This builder for chaining.
*/
public Builder setStatus(boolean value) {
status_ = value;
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>optional bool status = 1;</code>
* @return This builder for chaining.
*/
public Builder clearStatus() {
bitField0_ = (bitField0_ & ~0x00000001);
status_ = false;
onChanged();
return this;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.NamenodeHeartbeatResponseProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.NamenodeHeartbeatResponseProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatResponseProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatResponseProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatResponseProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<NamenodeHeartbeatResponseProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<NamenodeHeartbeatResponseProto>() {
@java.lang.Override
public NamenodeHeartbeatResponseProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<NamenodeHeartbeatResponseProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<NamenodeHeartbeatResponseProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.NamenodeHeartbeatResponseProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface RemoteLocationProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.RemoteLocationProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* <code>optional string nameserviceId = 1;</code>
* @return Whether the nameserviceId field is set.
*/
boolean hasNameserviceId();
/**
* <code>optional string nameserviceId = 1;</code>
* @return The nameserviceId.
*/
java.lang.String getNameserviceId();
/**
* <code>optional string nameserviceId = 1;</code>
* @return The bytes for nameserviceId.
*/
org.apache.hadoop.thirdparty.protobuf.ByteString
getNameserviceIdBytes();
/**
* <code>optional string path = 2;</code>
* @return Whether the path field is set.
*/
boolean hasPath();
/**
* <code>optional string path = 2;</code>
* @return The path.
*/
java.lang.String getPath();
/**
* <code>optional string path = 2;</code>
* @return The bytes for path.
*/
org.apache.hadoop.thirdparty.protobuf.ByteString
getPathBytes();
}
/**
* Protobuf type {@code hadoop.hdfs.RemoteLocationProto}
*/
public static final class RemoteLocationProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.RemoteLocationProto)
RemoteLocationProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use RemoteLocationProto.newBuilder() to construct.
private RemoteLocationProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private RemoteLocationProto() {
nameserviceId_ = "";
path_ = "";
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new RemoteLocationProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RemoteLocationProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RemoteLocationProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto.Builder.class);
}
private int bitField0_;
public static final int NAMESERVICEID_FIELD_NUMBER = 1;
@SuppressWarnings("serial")
private volatile java.lang.Object nameserviceId_ = "";
/**
* <code>optional string nameserviceId = 1;</code>
* @return Whether the nameserviceId field is set.
*/
@java.lang.Override
public boolean hasNameserviceId() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional string nameserviceId = 1;</code>
* @return The nameserviceId.
*/
@java.lang.Override
public java.lang.String getNameserviceId() {
java.lang.Object ref = nameserviceId_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
nameserviceId_ = s;
}
return s;
}
}
/**
* <code>optional string nameserviceId = 1;</code>
* @return The bytes for nameserviceId.
*/
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.ByteString
getNameserviceIdBytes() {
java.lang.Object ref = nameserviceId_;
if (ref instanceof java.lang.String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
nameserviceId_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
public static final int PATH_FIELD_NUMBER = 2;
@SuppressWarnings("serial")
private volatile java.lang.Object path_ = "";
/**
* <code>optional string path = 2;</code>
* @return Whether the path field is set.
*/
@java.lang.Override
public boolean hasPath() {
return ((bitField0_ & 0x00000002) != 0);
}
/**
* <code>optional string path = 2;</code>
* @return The path.
*/
@java.lang.Override
public java.lang.String getPath() {
java.lang.Object ref = path_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
path_ = s;
}
return s;
}
}
/**
* <code>optional string path = 2;</code>
* @return The bytes for path.
*/
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.ByteString
getPathBytes() {
java.lang.Object ref = path_;
if (ref instanceof java.lang.String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
path_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) != 0)) {
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 1, nameserviceId_);
}
if (((bitField0_ & 0x00000002) != 0)) {
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 2, path_);
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(1, nameserviceId_);
}
if (((bitField0_ & 0x00000002) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(2, path_);
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto) obj;
if (hasNameserviceId() != other.hasNameserviceId()) return false;
if (hasNameserviceId()) {
if (!getNameserviceId()
.equals(other.getNameserviceId())) return false;
}
if (hasPath() != other.hasPath()) return false;
if (hasPath()) {
if (!getPath()
.equals(other.getPath())) return false;
}
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (hasNameserviceId()) {
hash = (37 * hash) + NAMESERVICEID_FIELD_NUMBER;
hash = (53 * hash) + getNameserviceId().hashCode();
}
if (hasPath()) {
hash = (37 * hash) + PATH_FIELD_NUMBER;
hash = (53 * hash) + getPath().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.RemoteLocationProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.RemoteLocationProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RemoteLocationProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RemoteLocationProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
nameserviceId_ = "";
path_ = "";
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RemoteLocationProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto(this);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartial0(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto result) {
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) != 0)) {
result.nameserviceId_ = nameserviceId_;
to_bitField0_ |= 0x00000001;
}
if (((from_bitField0_ & 0x00000002) != 0)) {
result.path_ = path_;
to_bitField0_ |= 0x00000002;
}
result.bitField0_ |= to_bitField0_;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto.getDefaultInstance()) return this;
if (other.hasNameserviceId()) {
nameserviceId_ = other.nameserviceId_;
bitField0_ |= 0x00000001;
onChanged();
}
if (other.hasPath()) {
path_ = other.path_;
bitField0_ |= 0x00000002;
onChanged();
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 10: {
nameserviceId_ = input.readBytes();
bitField0_ |= 0x00000001;
break;
} // case 10
case 18: {
path_ = input.readBytes();
bitField0_ |= 0x00000002;
break;
} // case 18
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private java.lang.Object nameserviceId_ = "";
/**
* <code>optional string nameserviceId = 1;</code>
* @return Whether the nameserviceId field is set.
*/
public boolean hasNameserviceId() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional string nameserviceId = 1;</code>
* @return The nameserviceId.
*/
public java.lang.String getNameserviceId() {
java.lang.Object ref = nameserviceId_;
if (!(ref instanceof java.lang.String)) {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
nameserviceId_ = s;
}
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* <code>optional string nameserviceId = 1;</code>
* @return The bytes for nameserviceId.
*/
public org.apache.hadoop.thirdparty.protobuf.ByteString
getNameserviceIdBytes() {
java.lang.Object ref = nameserviceId_;
if (ref instanceof String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
nameserviceId_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
/**
* <code>optional string nameserviceId = 1;</code>
* @param value The nameserviceId to set.
* @return This builder for chaining.
*/
public Builder setNameserviceId(
java.lang.String value) {
if (value == null) { throw new NullPointerException(); }
nameserviceId_ = value;
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>optional string nameserviceId = 1;</code>
* @return This builder for chaining.
*/
public Builder clearNameserviceId() {
nameserviceId_ = getDefaultInstance().getNameserviceId();
bitField0_ = (bitField0_ & ~0x00000001);
onChanged();
return this;
}
/**
* <code>optional string nameserviceId = 1;</code>
* @param value The bytes for nameserviceId to set.
* @return This builder for chaining.
*/
public Builder setNameserviceIdBytes(
org.apache.hadoop.thirdparty.protobuf.ByteString value) {
if (value == null) { throw new NullPointerException(); }
nameserviceId_ = value;
bitField0_ |= 0x00000001;
onChanged();
return this;
}
private java.lang.Object path_ = "";
/**
* <code>optional string path = 2;</code>
* @return Whether the path field is set.
*/
public boolean hasPath() {
return ((bitField0_ & 0x00000002) != 0);
}
/**
* <code>optional string path = 2;</code>
* @return The path.
*/
public java.lang.String getPath() {
java.lang.Object ref = path_;
if (!(ref instanceof java.lang.String)) {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
path_ = s;
}
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* <code>optional string path = 2;</code>
* @return The bytes for path.
*/
public org.apache.hadoop.thirdparty.protobuf.ByteString
getPathBytes() {
java.lang.Object ref = path_;
if (ref instanceof String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
path_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
/**
* <code>optional string path = 2;</code>
* @param value The path to set.
* @return This builder for chaining.
*/
public Builder setPath(
java.lang.String value) {
if (value == null) { throw new NullPointerException(); }
path_ = value;
bitField0_ |= 0x00000002;
onChanged();
return this;
}
/**
* <code>optional string path = 2;</code>
* @return This builder for chaining.
*/
public Builder clearPath() {
path_ = getDefaultInstance().getPath();
bitField0_ = (bitField0_ & ~0x00000002);
onChanged();
return this;
}
/**
* <code>optional string path = 2;</code>
* @param value The bytes for path to set.
* @return This builder for chaining.
*/
public Builder setPathBytes(
org.apache.hadoop.thirdparty.protobuf.ByteString value) {
if (value == null) { throw new NullPointerException(); }
path_ = value;
bitField0_ |= 0x00000002;
onChanged();
return this;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.RemoteLocationProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.RemoteLocationProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<RemoteLocationProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<RemoteLocationProto>() {
@java.lang.Override
public RemoteLocationProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<RemoteLocationProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<RemoteLocationProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface MountTableRecordProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.MountTableRecordProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* <code>optional string srcPath = 1;</code>
* @return Whether the srcPath field is set.
*/
boolean hasSrcPath();
/**
* <code>optional string srcPath = 1;</code>
* @return The srcPath.
*/
java.lang.String getSrcPath();
/**
* <code>optional string srcPath = 1;</code>
* @return The bytes for srcPath.
*/
org.apache.hadoop.thirdparty.protobuf.ByteString
getSrcPathBytes();
/**
* <code>repeated .hadoop.hdfs.RemoteLocationProto destinations = 2;</code>
*/
java.util.List<org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto>
getDestinationsList();
/**
* <code>repeated .hadoop.hdfs.RemoteLocationProto destinations = 2;</code>
*/
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto getDestinations(int index);
/**
* <code>repeated .hadoop.hdfs.RemoteLocationProto destinations = 2;</code>
*/
int getDestinationsCount();
/**
* <code>repeated .hadoop.hdfs.RemoteLocationProto destinations = 2;</code>
*/
java.util.List<? extends org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProtoOrBuilder>
getDestinationsOrBuilderList();
/**
* <code>repeated .hadoop.hdfs.RemoteLocationProto destinations = 2;</code>
*/
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProtoOrBuilder getDestinationsOrBuilder(
int index);
/**
* <code>optional uint64 dateCreated = 3;</code>
* @return Whether the dateCreated field is set.
*/
boolean hasDateCreated();
/**
* <code>optional uint64 dateCreated = 3;</code>
* @return The dateCreated.
*/
long getDateCreated();
/**
* <code>optional uint64 dateModified = 4;</code>
* @return Whether the dateModified field is set.
*/
boolean hasDateModified();
/**
* <code>optional uint64 dateModified = 4;</code>
* @return The dateModified.
*/
long getDateModified();
/**
* <code>optional bool readOnly = 5 [default = false];</code>
* @return Whether the readOnly field is set.
*/
boolean hasReadOnly();
/**
* <code>optional bool readOnly = 5 [default = false];</code>
* @return The readOnly.
*/
boolean getReadOnly();
/**
* <code>optional .hadoop.hdfs.MountTableRecordProto.DestOrder destOrder = 6 [default = HASH];</code>
* @return Whether the destOrder field is set.
*/
boolean hasDestOrder();
/**
* <code>optional .hadoop.hdfs.MountTableRecordProto.DestOrder destOrder = 6 [default = HASH];</code>
* @return The destOrder.
*/
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.DestOrder getDestOrder();
/**
* <code>optional string ownerName = 10;</code>
* @return Whether the ownerName field is set.
*/
boolean hasOwnerName();
/**
* <code>optional string ownerName = 10;</code>
* @return The ownerName.
*/
java.lang.String getOwnerName();
/**
* <code>optional string ownerName = 10;</code>
* @return The bytes for ownerName.
*/
org.apache.hadoop.thirdparty.protobuf.ByteString
getOwnerNameBytes();
/**
* <code>optional string groupName = 11;</code>
* @return Whether the groupName field is set.
*/
boolean hasGroupName();
/**
* <code>optional string groupName = 11;</code>
* @return The groupName.
*/
java.lang.String getGroupName();
/**
* <code>optional string groupName = 11;</code>
* @return The bytes for groupName.
*/
org.apache.hadoop.thirdparty.protobuf.ByteString
getGroupNameBytes();
/**
* <code>optional int32 mode = 12;</code>
* @return Whether the mode field is set.
*/
boolean hasMode();
/**
* <code>optional int32 mode = 12;</code>
* @return The mode.
*/
int getMode();
/**
* <code>optional .hadoop.hdfs.QuotaUsageProto quota = 13;</code>
* @return Whether the quota field is set.
*/
boolean hasQuota();
/**
* <code>optional .hadoop.hdfs.QuotaUsageProto quota = 13;</code>
* @return The quota.
*/
org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.QuotaUsageProto getQuota();
/**
* <code>optional .hadoop.hdfs.QuotaUsageProto quota = 13;</code>
*/
org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.QuotaUsageProtoOrBuilder getQuotaOrBuilder();
/**
* <code>optional bool faultTolerant = 14 [default = false];</code>
* @return Whether the faultTolerant field is set.
*/
boolean hasFaultTolerant();
/**
* <code>optional bool faultTolerant = 14 [default = false];</code>
* @return The faultTolerant.
*/
boolean getFaultTolerant();
}
/**
* Protobuf type {@code hadoop.hdfs.MountTableRecordProto}
*/
public static final class MountTableRecordProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.MountTableRecordProto)
MountTableRecordProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use MountTableRecordProto.newBuilder() to construct.
private MountTableRecordProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private MountTableRecordProto() {
srcPath_ = "";
destinations_ = java.util.Collections.emptyList();
destOrder_ = 0;
ownerName_ = "";
groupName_ = "";
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new MountTableRecordProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_MountTableRecordProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_MountTableRecordProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.Builder.class);
}
/**
* Protobuf enum {@code hadoop.hdfs.MountTableRecordProto.DestOrder}
*/
public enum DestOrder
implements org.apache.hadoop.thirdparty.protobuf.ProtocolMessageEnum {
/**
* <code>HASH = 0;</code>
*/
HASH(0),
/**
* <code>LOCAL = 1;</code>
*/
LOCAL(1),
/**
* <code>RANDOM = 2;</code>
*/
RANDOM(2),
/**
* <code>HASH_ALL = 3;</code>
*/
HASH_ALL(3),
/**
* <code>SPACE = 4;</code>
*/
SPACE(4),
/**
* <code>LEADER_FOLLOWER = 5;</code>
*/
LEADER_FOLLOWER(5),
;
/**
* <code>HASH = 0;</code>
*/
public static final int HASH_VALUE = 0;
/**
* <code>LOCAL = 1;</code>
*/
public static final int LOCAL_VALUE = 1;
/**
* <code>RANDOM = 2;</code>
*/
public static final int RANDOM_VALUE = 2;
/**
* <code>HASH_ALL = 3;</code>
*/
public static final int HASH_ALL_VALUE = 3;
/**
* <code>SPACE = 4;</code>
*/
public static final int SPACE_VALUE = 4;
/**
* <code>LEADER_FOLLOWER = 5;</code>
*/
public static final int LEADER_FOLLOWER_VALUE = 5;
public final int getNumber() {
return value;
}
/**
* @param value The numeric wire value of the corresponding enum entry.
* @return The enum associated with the given numeric wire value.
* @deprecated Use {@link #forNumber(int)} instead.
*/
@java.lang.Deprecated
public static DestOrder valueOf(int value) {
return forNumber(value);
}
/**
* @param value The numeric wire value of the corresponding enum entry.
* @return The enum associated with the given numeric wire value.
*/
public static DestOrder forNumber(int value) {
switch (value) {
case 0: return HASH;
case 1: return LOCAL;
case 2: return RANDOM;
case 3: return HASH_ALL;
case 4: return SPACE;
case 5: return LEADER_FOLLOWER;
default: return null;
}
}
public static org.apache.hadoop.thirdparty.protobuf.Internal.EnumLiteMap<DestOrder>
internalGetValueMap() {
return internalValueMap;
}
private static final org.apache.hadoop.thirdparty.protobuf.Internal.EnumLiteMap<
DestOrder> internalValueMap =
new org.apache.hadoop.thirdparty.protobuf.Internal.EnumLiteMap<DestOrder>() {
public DestOrder findValueByNumber(int number) {
return DestOrder.forNumber(number);
}
};
public final org.apache.hadoop.thirdparty.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
return getDescriptor().getValues().get(ordinal());
}
public final org.apache.hadoop.thirdparty.protobuf.Descriptors.EnumDescriptor
getDescriptorForType() {
return getDescriptor();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.EnumDescriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.getDescriptor().getEnumTypes().get(0);
}
private static final DestOrder[] VALUES = values();
public static DestOrder valueOf(
org.apache.hadoop.thirdparty.protobuf.Descriptors.EnumValueDescriptor desc) {
if (desc.getType() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"EnumValueDescriptor is not for this type.");
}
return VALUES[desc.getIndex()];
}
private final int value;
private DestOrder(int value) {
this.value = value;
}
// @@protoc_insertion_point(enum_scope:hadoop.hdfs.MountTableRecordProto.DestOrder)
}
private int bitField0_;
public static final int SRCPATH_FIELD_NUMBER = 1;
@SuppressWarnings("serial")
private volatile java.lang.Object srcPath_ = "";
/**
* <code>optional string srcPath = 1;</code>
* @return Whether the srcPath field is set.
*/
@java.lang.Override
public boolean hasSrcPath() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional string srcPath = 1;</code>
* @return The srcPath.
*/
@java.lang.Override
public java.lang.String getSrcPath() {
java.lang.Object ref = srcPath_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
srcPath_ = s;
}
return s;
}
}
/**
* <code>optional string srcPath = 1;</code>
* @return The bytes for srcPath.
*/
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.ByteString
getSrcPathBytes() {
java.lang.Object ref = srcPath_;
if (ref instanceof java.lang.String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
srcPath_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
public static final int DESTINATIONS_FIELD_NUMBER = 2;
@SuppressWarnings("serial")
private java.util.List<org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto> destinations_;
/**
* <code>repeated .hadoop.hdfs.RemoteLocationProto destinations = 2;</code>
*/
@java.lang.Override
public java.util.List<org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto> getDestinationsList() {
return destinations_;
}
/**
* <code>repeated .hadoop.hdfs.RemoteLocationProto destinations = 2;</code>
*/
@java.lang.Override
public java.util.List<? extends org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProtoOrBuilder>
getDestinationsOrBuilderList() {
return destinations_;
}
/**
* <code>repeated .hadoop.hdfs.RemoteLocationProto destinations = 2;</code>
*/
@java.lang.Override
public int getDestinationsCount() {
return destinations_.size();
}
/**
* <code>repeated .hadoop.hdfs.RemoteLocationProto destinations = 2;</code>
*/
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto getDestinations(int index) {
return destinations_.get(index);
}
/**
* <code>repeated .hadoop.hdfs.RemoteLocationProto destinations = 2;</code>
*/
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProtoOrBuilder getDestinationsOrBuilder(
int index) {
return destinations_.get(index);
}
public static final int DATECREATED_FIELD_NUMBER = 3;
private long dateCreated_ = 0L;
/**
* <code>optional uint64 dateCreated = 3;</code>
* @return Whether the dateCreated field is set.
*/
@java.lang.Override
public boolean hasDateCreated() {
return ((bitField0_ & 0x00000002) != 0);
}
/**
* <code>optional uint64 dateCreated = 3;</code>
* @return The dateCreated.
*/
@java.lang.Override
public long getDateCreated() {
return dateCreated_;
}
public static final int DATEMODIFIED_FIELD_NUMBER = 4;
private long dateModified_ = 0L;
/**
* <code>optional uint64 dateModified = 4;</code>
* @return Whether the dateModified field is set.
*/
@java.lang.Override
public boolean hasDateModified() {
return ((bitField0_ & 0x00000004) != 0);
}
/**
* <code>optional uint64 dateModified = 4;</code>
* @return The dateModified.
*/
@java.lang.Override
public long getDateModified() {
return dateModified_;
}
public static final int READONLY_FIELD_NUMBER = 5;
private boolean readOnly_ = false;
/**
* <code>optional bool readOnly = 5 [default = false];</code>
* @return Whether the readOnly field is set.
*/
@java.lang.Override
public boolean hasReadOnly() {
return ((bitField0_ & 0x00000008) != 0);
}
/**
* <code>optional bool readOnly = 5 [default = false];</code>
* @return The readOnly.
*/
@java.lang.Override
public boolean getReadOnly() {
return readOnly_;
}
public static final int DESTORDER_FIELD_NUMBER = 6;
private int destOrder_ = 0;
/**
* <code>optional .hadoop.hdfs.MountTableRecordProto.DestOrder destOrder = 6 [default = HASH];</code>
* @return Whether the destOrder field is set.
*/
@java.lang.Override public boolean hasDestOrder() {
return ((bitField0_ & 0x00000010) != 0);
}
/**
* <code>optional .hadoop.hdfs.MountTableRecordProto.DestOrder destOrder = 6 [default = HASH];</code>
* @return The destOrder.
*/
@java.lang.Override public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.DestOrder getDestOrder() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.DestOrder result = org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.DestOrder.forNumber(destOrder_);
return result == null ? org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.DestOrder.HASH : result;
}
public static final int OWNERNAME_FIELD_NUMBER = 10;
@SuppressWarnings("serial")
private volatile java.lang.Object ownerName_ = "";
/**
* <code>optional string ownerName = 10;</code>
* @return Whether the ownerName field is set.
*/
@java.lang.Override
public boolean hasOwnerName() {
return ((bitField0_ & 0x00000020) != 0);
}
/**
* <code>optional string ownerName = 10;</code>
* @return The ownerName.
*/
@java.lang.Override
public java.lang.String getOwnerName() {
java.lang.Object ref = ownerName_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
ownerName_ = s;
}
return s;
}
}
/**
* <code>optional string ownerName = 10;</code>
* @return The bytes for ownerName.
*/
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.ByteString
getOwnerNameBytes() {
java.lang.Object ref = ownerName_;
if (ref instanceof java.lang.String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
ownerName_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
public static final int GROUPNAME_FIELD_NUMBER = 11;
@SuppressWarnings("serial")
private volatile java.lang.Object groupName_ = "";
/**
* <code>optional string groupName = 11;</code>
* @return Whether the groupName field is set.
*/
@java.lang.Override
public boolean hasGroupName() {
return ((bitField0_ & 0x00000040) != 0);
}
/**
* <code>optional string groupName = 11;</code>
* @return The groupName.
*/
@java.lang.Override
public java.lang.String getGroupName() {
java.lang.Object ref = groupName_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
groupName_ = s;
}
return s;
}
}
/**
* <code>optional string groupName = 11;</code>
* @return The bytes for groupName.
*/
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.ByteString
getGroupNameBytes() {
java.lang.Object ref = groupName_;
if (ref instanceof java.lang.String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
groupName_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
public static final int MODE_FIELD_NUMBER = 12;
private int mode_ = 0;
/**
* <code>optional int32 mode = 12;</code>
* @return Whether the mode field is set.
*/
@java.lang.Override
public boolean hasMode() {
return ((bitField0_ & 0x00000080) != 0);
}
/**
* <code>optional int32 mode = 12;</code>
* @return The mode.
*/
@java.lang.Override
public int getMode() {
return mode_;
}
public static final int QUOTA_FIELD_NUMBER = 13;
private org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.QuotaUsageProto quota_;
/**
* <code>optional .hadoop.hdfs.QuotaUsageProto quota = 13;</code>
* @return Whether the quota field is set.
*/
@java.lang.Override
public boolean hasQuota() {
return ((bitField0_ & 0x00000100) != 0);
}
/**
* <code>optional .hadoop.hdfs.QuotaUsageProto quota = 13;</code>
* @return The quota.
*/
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.QuotaUsageProto getQuota() {
return quota_ == null ? org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.QuotaUsageProto.getDefaultInstance() : quota_;
}
/**
* <code>optional .hadoop.hdfs.QuotaUsageProto quota = 13;</code>
*/
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.QuotaUsageProtoOrBuilder getQuotaOrBuilder() {
return quota_ == null ? org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.QuotaUsageProto.getDefaultInstance() : quota_;
}
public static final int FAULTTOLERANT_FIELD_NUMBER = 14;
private boolean faultTolerant_ = false;
/**
* <code>optional bool faultTolerant = 14 [default = false];</code>
* @return Whether the faultTolerant field is set.
*/
@java.lang.Override
public boolean hasFaultTolerant() {
return ((bitField0_ & 0x00000200) != 0);
}
/**
* <code>optional bool faultTolerant = 14 [default = false];</code>
* @return The faultTolerant.
*/
@java.lang.Override
public boolean getFaultTolerant() {
return faultTolerant_;
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
if (hasQuota()) {
if (!getQuota().isInitialized()) {
memoizedIsInitialized = 0;
return false;
}
}
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) != 0)) {
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 1, srcPath_);
}
for (int i = 0; i < destinations_.size(); i++) {
output.writeMessage(2, destinations_.get(i));
}
if (((bitField0_ & 0x00000002) != 0)) {
output.writeUInt64(3, dateCreated_);
}
if (((bitField0_ & 0x00000004) != 0)) {
output.writeUInt64(4, dateModified_);
}
if (((bitField0_ & 0x00000008) != 0)) {
output.writeBool(5, readOnly_);
}
if (((bitField0_ & 0x00000010) != 0)) {
output.writeEnum(6, destOrder_);
}
if (((bitField0_ & 0x00000020) != 0)) {
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 10, ownerName_);
}
if (((bitField0_ & 0x00000040) != 0)) {
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 11, groupName_);
}
if (((bitField0_ & 0x00000080) != 0)) {
output.writeInt32(12, mode_);
}
if (((bitField0_ & 0x00000100) != 0)) {
output.writeMessage(13, getQuota());
}
if (((bitField0_ & 0x00000200) != 0)) {
output.writeBool(14, faultTolerant_);
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(1, srcPath_);
}
for (int i = 0; i < destinations_.size(); i++) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeMessageSize(2, destinations_.get(i));
}
if (((bitField0_ & 0x00000002) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeUInt64Size(3, dateCreated_);
}
if (((bitField0_ & 0x00000004) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeUInt64Size(4, dateModified_);
}
if (((bitField0_ & 0x00000008) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeBoolSize(5, readOnly_);
}
if (((bitField0_ & 0x00000010) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeEnumSize(6, destOrder_);
}
if (((bitField0_ & 0x00000020) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(10, ownerName_);
}
if (((bitField0_ & 0x00000040) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(11, groupName_);
}
if (((bitField0_ & 0x00000080) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeInt32Size(12, mode_);
}
if (((bitField0_ & 0x00000100) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeMessageSize(13, getQuota());
}
if (((bitField0_ & 0x00000200) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeBoolSize(14, faultTolerant_);
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto) obj;
if (hasSrcPath() != other.hasSrcPath()) return false;
if (hasSrcPath()) {
if (!getSrcPath()
.equals(other.getSrcPath())) return false;
}
if (!getDestinationsList()
.equals(other.getDestinationsList())) return false;
if (hasDateCreated() != other.hasDateCreated()) return false;
if (hasDateCreated()) {
if (getDateCreated()
!= other.getDateCreated()) return false;
}
if (hasDateModified() != other.hasDateModified()) return false;
if (hasDateModified()) {
if (getDateModified()
!= other.getDateModified()) return false;
}
if (hasReadOnly() != other.hasReadOnly()) return false;
if (hasReadOnly()) {
if (getReadOnly()
!= other.getReadOnly()) return false;
}
if (hasDestOrder() != other.hasDestOrder()) return false;
if (hasDestOrder()) {
if (destOrder_ != other.destOrder_) return false;
}
if (hasOwnerName() != other.hasOwnerName()) return false;
if (hasOwnerName()) {
if (!getOwnerName()
.equals(other.getOwnerName())) return false;
}
if (hasGroupName() != other.hasGroupName()) return false;
if (hasGroupName()) {
if (!getGroupName()
.equals(other.getGroupName())) return false;
}
if (hasMode() != other.hasMode()) return false;
if (hasMode()) {
if (getMode()
!= other.getMode()) return false;
}
if (hasQuota() != other.hasQuota()) return false;
if (hasQuota()) {
if (!getQuota()
.equals(other.getQuota())) return false;
}
if (hasFaultTolerant() != other.hasFaultTolerant()) return false;
if (hasFaultTolerant()) {
if (getFaultTolerant()
!= other.getFaultTolerant()) return false;
}
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (hasSrcPath()) {
hash = (37 * hash) + SRCPATH_FIELD_NUMBER;
hash = (53 * hash) + getSrcPath().hashCode();
}
if (getDestinationsCount() > 0) {
hash = (37 * hash) + DESTINATIONS_FIELD_NUMBER;
hash = (53 * hash) + getDestinationsList().hashCode();
}
if (hasDateCreated()) {
hash = (37 * hash) + DATECREATED_FIELD_NUMBER;
hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashLong(
getDateCreated());
}
if (hasDateModified()) {
hash = (37 * hash) + DATEMODIFIED_FIELD_NUMBER;
hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashLong(
getDateModified());
}
if (hasReadOnly()) {
hash = (37 * hash) + READONLY_FIELD_NUMBER;
hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashBoolean(
getReadOnly());
}
if (hasDestOrder()) {
hash = (37 * hash) + DESTORDER_FIELD_NUMBER;
hash = (53 * hash) + destOrder_;
}
if (hasOwnerName()) {
hash = (37 * hash) + OWNERNAME_FIELD_NUMBER;
hash = (53 * hash) + getOwnerName().hashCode();
}
if (hasGroupName()) {
hash = (37 * hash) + GROUPNAME_FIELD_NUMBER;
hash = (53 * hash) + getGroupName().hashCode();
}
if (hasMode()) {
hash = (37 * hash) + MODE_FIELD_NUMBER;
hash = (53 * hash) + getMode();
}
if (hasQuota()) {
hash = (37 * hash) + QUOTA_FIELD_NUMBER;
hash = (53 * hash) + getQuota().hashCode();
}
if (hasFaultTolerant()) {
hash = (37 * hash) + FAULTTOLERANT_FIELD_NUMBER;
hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashBoolean(
getFaultTolerant());
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.MountTableRecordProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.MountTableRecordProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_MountTableRecordProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_MountTableRecordProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
if (org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
getDestinationsFieldBuilder();
getQuotaFieldBuilder();
}
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
srcPath_ = "";
if (destinationsBuilder_ == null) {
destinations_ = java.util.Collections.emptyList();
} else {
destinations_ = null;
destinationsBuilder_.clear();
}
bitField0_ = (bitField0_ & ~0x00000002);
dateCreated_ = 0L;
dateModified_ = 0L;
readOnly_ = false;
destOrder_ = 0;
ownerName_ = "";
groupName_ = "";
mode_ = 0;
quota_ = null;
if (quotaBuilder_ != null) {
quotaBuilder_.dispose();
quotaBuilder_ = null;
}
faultTolerant_ = false;
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_MountTableRecordProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto(this);
buildPartialRepeatedFields(result);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartialRepeatedFields(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto result) {
if (destinationsBuilder_ == null) {
if (((bitField0_ & 0x00000002) != 0)) {
destinations_ = java.util.Collections.unmodifiableList(destinations_);
bitField0_ = (bitField0_ & ~0x00000002);
}
result.destinations_ = destinations_;
} else {
result.destinations_ = destinationsBuilder_.build();
}
}
private void buildPartial0(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto result) {
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) != 0)) {
result.srcPath_ = srcPath_;
to_bitField0_ |= 0x00000001;
}
if (((from_bitField0_ & 0x00000004) != 0)) {
result.dateCreated_ = dateCreated_;
to_bitField0_ |= 0x00000002;
}
if (((from_bitField0_ & 0x00000008) != 0)) {
result.dateModified_ = dateModified_;
to_bitField0_ |= 0x00000004;
}
if (((from_bitField0_ & 0x00000010) != 0)) {
result.readOnly_ = readOnly_;
to_bitField0_ |= 0x00000008;
}
if (((from_bitField0_ & 0x00000020) != 0)) {
result.destOrder_ = destOrder_;
to_bitField0_ |= 0x00000010;
}
if (((from_bitField0_ & 0x00000040) != 0)) {
result.ownerName_ = ownerName_;
to_bitField0_ |= 0x00000020;
}
if (((from_bitField0_ & 0x00000080) != 0)) {
result.groupName_ = groupName_;
to_bitField0_ |= 0x00000040;
}
if (((from_bitField0_ & 0x00000100) != 0)) {
result.mode_ = mode_;
to_bitField0_ |= 0x00000080;
}
if (((from_bitField0_ & 0x00000200) != 0)) {
result.quota_ = quotaBuilder_ == null
? quota_
: quotaBuilder_.build();
to_bitField0_ |= 0x00000100;
}
if (((from_bitField0_ & 0x00000400) != 0)) {
result.faultTolerant_ = faultTolerant_;
to_bitField0_ |= 0x00000200;
}
result.bitField0_ |= to_bitField0_;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.getDefaultInstance()) return this;
if (other.hasSrcPath()) {
srcPath_ = other.srcPath_;
bitField0_ |= 0x00000001;
onChanged();
}
if (destinationsBuilder_ == null) {
if (!other.destinations_.isEmpty()) {
if (destinations_.isEmpty()) {
destinations_ = other.destinations_;
bitField0_ = (bitField0_ & ~0x00000002);
} else {
ensureDestinationsIsMutable();
destinations_.addAll(other.destinations_);
}
onChanged();
}
} else {
if (!other.destinations_.isEmpty()) {
if (destinationsBuilder_.isEmpty()) {
destinationsBuilder_.dispose();
destinationsBuilder_ = null;
destinations_ = other.destinations_;
bitField0_ = (bitField0_ & ~0x00000002);
destinationsBuilder_ =
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
getDestinationsFieldBuilder() : null;
} else {
destinationsBuilder_.addAllMessages(other.destinations_);
}
}
}
if (other.hasDateCreated()) {
setDateCreated(other.getDateCreated());
}
if (other.hasDateModified()) {
setDateModified(other.getDateModified());
}
if (other.hasReadOnly()) {
setReadOnly(other.getReadOnly());
}
if (other.hasDestOrder()) {
setDestOrder(other.getDestOrder());
}
if (other.hasOwnerName()) {
ownerName_ = other.ownerName_;
bitField0_ |= 0x00000040;
onChanged();
}
if (other.hasGroupName()) {
groupName_ = other.groupName_;
bitField0_ |= 0x00000080;
onChanged();
}
if (other.hasMode()) {
setMode(other.getMode());
}
if (other.hasQuota()) {
mergeQuota(other.getQuota());
}
if (other.hasFaultTolerant()) {
setFaultTolerant(other.getFaultTolerant());
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
if (hasQuota()) {
if (!getQuota().isInitialized()) {
return false;
}
}
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 10: {
srcPath_ = input.readBytes();
bitField0_ |= 0x00000001;
break;
} // case 10
case 18: {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto m =
input.readMessage(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto.PARSER,
extensionRegistry);
if (destinationsBuilder_ == null) {
ensureDestinationsIsMutable();
destinations_.add(m);
} else {
destinationsBuilder_.addMessage(m);
}
break;
} // case 18
case 24: {
dateCreated_ = input.readUInt64();
bitField0_ |= 0x00000004;
break;
} // case 24
case 32: {
dateModified_ = input.readUInt64();
bitField0_ |= 0x00000008;
break;
} // case 32
case 40: {
readOnly_ = input.readBool();
bitField0_ |= 0x00000010;
break;
} // case 40
case 48: {
int tmpRaw = input.readEnum();
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.DestOrder tmpValue =
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.DestOrder.forNumber(tmpRaw);
if (tmpValue == null) {
mergeUnknownVarintField(6, tmpRaw);
} else {
destOrder_ = tmpRaw;
bitField0_ |= 0x00000020;
}
break;
} // case 48
case 82: {
ownerName_ = input.readBytes();
bitField0_ |= 0x00000040;
break;
} // case 82
case 90: {
groupName_ = input.readBytes();
bitField0_ |= 0x00000080;
break;
} // case 90
case 96: {
mode_ = input.readInt32();
bitField0_ |= 0x00000100;
break;
} // case 96
case 106: {
input.readMessage(
getQuotaFieldBuilder().getBuilder(),
extensionRegistry);
bitField0_ |= 0x00000200;
break;
} // case 106
case 112: {
faultTolerant_ = input.readBool();
bitField0_ |= 0x00000400;
break;
} // case 112
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private java.lang.Object srcPath_ = "";
/**
* <code>optional string srcPath = 1;</code>
* @return Whether the srcPath field is set.
*/
public boolean hasSrcPath() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional string srcPath = 1;</code>
* @return The srcPath.
*/
public java.lang.String getSrcPath() {
java.lang.Object ref = srcPath_;
if (!(ref instanceof java.lang.String)) {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
srcPath_ = s;
}
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* <code>optional string srcPath = 1;</code>
* @return The bytes for srcPath.
*/
public org.apache.hadoop.thirdparty.protobuf.ByteString
getSrcPathBytes() {
java.lang.Object ref = srcPath_;
if (ref instanceof String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
srcPath_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
/**
* <code>optional string srcPath = 1;</code>
* @param value The srcPath to set.
* @return This builder for chaining.
*/
public Builder setSrcPath(
java.lang.String value) {
if (value == null) { throw new NullPointerException(); }
srcPath_ = value;
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>optional string srcPath = 1;</code>
* @return This builder for chaining.
*/
public Builder clearSrcPath() {
srcPath_ = getDefaultInstance().getSrcPath();
bitField0_ = (bitField0_ & ~0x00000001);
onChanged();
return this;
}
/**
* <code>optional string srcPath = 1;</code>
* @param value The bytes for srcPath to set.
* @return This builder for chaining.
*/
public Builder setSrcPathBytes(
org.apache.hadoop.thirdparty.protobuf.ByteString value) {
if (value == null) { throw new NullPointerException(); }
srcPath_ = value;
bitField0_ |= 0x00000001;
onChanged();
return this;
}
private java.util.List<org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto> destinations_ =
java.util.Collections.emptyList();
private void ensureDestinationsIsMutable() {
if (!((bitField0_ & 0x00000002) != 0)) {
destinations_ = new java.util.ArrayList<org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto>(destinations_);
bitField0_ |= 0x00000002;
}
}
private org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto.Builder, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProtoOrBuilder> destinationsBuilder_;
/**
* <code>repeated .hadoop.hdfs.RemoteLocationProto destinations = 2;</code>
*/
public java.util.List<org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto> getDestinationsList() {
if (destinationsBuilder_ == null) {
return java.util.Collections.unmodifiableList(destinations_);
} else {
return destinationsBuilder_.getMessageList();
}
}
/**
* <code>repeated .hadoop.hdfs.RemoteLocationProto destinations = 2;</code>
*/
public int getDestinationsCount() {
if (destinationsBuilder_ == null) {
return destinations_.size();
} else {
return destinationsBuilder_.getCount();
}
}
/**
* <code>repeated .hadoop.hdfs.RemoteLocationProto destinations = 2;</code>
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto getDestinations(int index) {
if (destinationsBuilder_ == null) {
return destinations_.get(index);
} else {
return destinationsBuilder_.getMessage(index);
}
}
/**
* <code>repeated .hadoop.hdfs.RemoteLocationProto destinations = 2;</code>
*/
public Builder setDestinations(
int index, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto value) {
if (destinationsBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
ensureDestinationsIsMutable();
destinations_.set(index, value);
onChanged();
} else {
destinationsBuilder_.setMessage(index, value);
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.RemoteLocationProto destinations = 2;</code>
*/
public Builder setDestinations(
int index, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto.Builder builderForValue) {
if (destinationsBuilder_ == null) {
ensureDestinationsIsMutable();
destinations_.set(index, builderForValue.build());
onChanged();
} else {
destinationsBuilder_.setMessage(index, builderForValue.build());
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.RemoteLocationProto destinations = 2;</code>
*/
public Builder addDestinations(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto value) {
if (destinationsBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
ensureDestinationsIsMutable();
destinations_.add(value);
onChanged();
} else {
destinationsBuilder_.addMessage(value);
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.RemoteLocationProto destinations = 2;</code>
*/
public Builder addDestinations(
int index, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto value) {
if (destinationsBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
ensureDestinationsIsMutable();
destinations_.add(index, value);
onChanged();
} else {
destinationsBuilder_.addMessage(index, value);
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.RemoteLocationProto destinations = 2;</code>
*/
public Builder addDestinations(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto.Builder builderForValue) {
if (destinationsBuilder_ == null) {
ensureDestinationsIsMutable();
destinations_.add(builderForValue.build());
onChanged();
} else {
destinationsBuilder_.addMessage(builderForValue.build());
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.RemoteLocationProto destinations = 2;</code>
*/
public Builder addDestinations(
int index, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto.Builder builderForValue) {
if (destinationsBuilder_ == null) {
ensureDestinationsIsMutable();
destinations_.add(index, builderForValue.build());
onChanged();
} else {
destinationsBuilder_.addMessage(index, builderForValue.build());
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.RemoteLocationProto destinations = 2;</code>
*/
public Builder addAllDestinations(
java.lang.Iterable<? extends org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto> values) {
if (destinationsBuilder_ == null) {
ensureDestinationsIsMutable();
org.apache.hadoop.thirdparty.protobuf.AbstractMessageLite.Builder.addAll(
values, destinations_);
onChanged();
} else {
destinationsBuilder_.addAllMessages(values);
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.RemoteLocationProto destinations = 2;</code>
*/
public Builder clearDestinations() {
if (destinationsBuilder_ == null) {
destinations_ = java.util.Collections.emptyList();
bitField0_ = (bitField0_ & ~0x00000002);
onChanged();
} else {
destinationsBuilder_.clear();
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.RemoteLocationProto destinations = 2;</code>
*/
public Builder removeDestinations(int index) {
if (destinationsBuilder_ == null) {
ensureDestinationsIsMutable();
destinations_.remove(index);
onChanged();
} else {
destinationsBuilder_.remove(index);
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.RemoteLocationProto destinations = 2;</code>
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto.Builder getDestinationsBuilder(
int index) {
return getDestinationsFieldBuilder().getBuilder(index);
}
/**
* <code>repeated .hadoop.hdfs.RemoteLocationProto destinations = 2;</code>
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProtoOrBuilder getDestinationsOrBuilder(
int index) {
if (destinationsBuilder_ == null) {
return destinations_.get(index); } else {
return destinationsBuilder_.getMessageOrBuilder(index);
}
}
/**
* <code>repeated .hadoop.hdfs.RemoteLocationProto destinations = 2;</code>
*/
public java.util.List<? extends org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProtoOrBuilder>
getDestinationsOrBuilderList() {
if (destinationsBuilder_ != null) {
return destinationsBuilder_.getMessageOrBuilderList();
} else {
return java.util.Collections.unmodifiableList(destinations_);
}
}
/**
* <code>repeated .hadoop.hdfs.RemoteLocationProto destinations = 2;</code>
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto.Builder addDestinationsBuilder() {
return getDestinationsFieldBuilder().addBuilder(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto.getDefaultInstance());
}
/**
* <code>repeated .hadoop.hdfs.RemoteLocationProto destinations = 2;</code>
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto.Builder addDestinationsBuilder(
int index) {
return getDestinationsFieldBuilder().addBuilder(
index, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto.getDefaultInstance());
}
/**
* <code>repeated .hadoop.hdfs.RemoteLocationProto destinations = 2;</code>
*/
public java.util.List<org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto.Builder>
getDestinationsBuilderList() {
return getDestinationsFieldBuilder().getBuilderList();
}
private org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto.Builder, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProtoOrBuilder>
getDestinationsFieldBuilder() {
if (destinationsBuilder_ == null) {
destinationsBuilder_ = new org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProto.Builder, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoteLocationProtoOrBuilder>(
destinations_,
((bitField0_ & 0x00000002) != 0),
getParentForChildren(),
isClean());
destinations_ = null;
}
return destinationsBuilder_;
}
private long dateCreated_ ;
/**
* <code>optional uint64 dateCreated = 3;</code>
* @return Whether the dateCreated field is set.
*/
@java.lang.Override
public boolean hasDateCreated() {
return ((bitField0_ & 0x00000004) != 0);
}
/**
* <code>optional uint64 dateCreated = 3;</code>
* @return The dateCreated.
*/
@java.lang.Override
public long getDateCreated() {
return dateCreated_;
}
/**
* <code>optional uint64 dateCreated = 3;</code>
* @param value The dateCreated to set.
* @return This builder for chaining.
*/
public Builder setDateCreated(long value) {
dateCreated_ = value;
bitField0_ |= 0x00000004;
onChanged();
return this;
}
/**
* <code>optional uint64 dateCreated = 3;</code>
* @return This builder for chaining.
*/
public Builder clearDateCreated() {
bitField0_ = (bitField0_ & ~0x00000004);
dateCreated_ = 0L;
onChanged();
return this;
}
private long dateModified_ ;
/**
* <code>optional uint64 dateModified = 4;</code>
* @return Whether the dateModified field is set.
*/
@java.lang.Override
public boolean hasDateModified() {
return ((bitField0_ & 0x00000008) != 0);
}
/**
* <code>optional uint64 dateModified = 4;</code>
* @return The dateModified.
*/
@java.lang.Override
public long getDateModified() {
return dateModified_;
}
/**
* <code>optional uint64 dateModified = 4;</code>
* @param value The dateModified to set.
* @return This builder for chaining.
*/
public Builder setDateModified(long value) {
dateModified_ = value;
bitField0_ |= 0x00000008;
onChanged();
return this;
}
/**
* <code>optional uint64 dateModified = 4;</code>
* @return This builder for chaining.
*/
public Builder clearDateModified() {
bitField0_ = (bitField0_ & ~0x00000008);
dateModified_ = 0L;
onChanged();
return this;
}
private boolean readOnly_ ;
/**
* <code>optional bool readOnly = 5 [default = false];</code>
* @return Whether the readOnly field is set.
*/
@java.lang.Override
public boolean hasReadOnly() {
return ((bitField0_ & 0x00000010) != 0);
}
/**
* <code>optional bool readOnly = 5 [default = false];</code>
* @return The readOnly.
*/
@java.lang.Override
public boolean getReadOnly() {
return readOnly_;
}
/**
* <code>optional bool readOnly = 5 [default = false];</code>
* @param value The readOnly to set.
* @return This builder for chaining.
*/
public Builder setReadOnly(boolean value) {
readOnly_ = value;
bitField0_ |= 0x00000010;
onChanged();
return this;
}
/**
* <code>optional bool readOnly = 5 [default = false];</code>
* @return This builder for chaining.
*/
public Builder clearReadOnly() {
bitField0_ = (bitField0_ & ~0x00000010);
readOnly_ = false;
onChanged();
return this;
}
private int destOrder_ = 0;
/**
* <code>optional .hadoop.hdfs.MountTableRecordProto.DestOrder destOrder = 6 [default = HASH];</code>
* @return Whether the destOrder field is set.
*/
@java.lang.Override public boolean hasDestOrder() {
return ((bitField0_ & 0x00000020) != 0);
}
/**
* <code>optional .hadoop.hdfs.MountTableRecordProto.DestOrder destOrder = 6 [default = HASH];</code>
* @return The destOrder.
*/
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.DestOrder getDestOrder() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.DestOrder result = org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.DestOrder.forNumber(destOrder_);
return result == null ? org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.DestOrder.HASH : result;
}
/**
* <code>optional .hadoop.hdfs.MountTableRecordProto.DestOrder destOrder = 6 [default = HASH];</code>
* @param value The destOrder to set.
* @return This builder for chaining.
*/
public Builder setDestOrder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.DestOrder value) {
if (value == null) {
throw new NullPointerException();
}
bitField0_ |= 0x00000020;
destOrder_ = value.getNumber();
onChanged();
return this;
}
/**
* <code>optional .hadoop.hdfs.MountTableRecordProto.DestOrder destOrder = 6 [default = HASH];</code>
* @return This builder for chaining.
*/
public Builder clearDestOrder() {
bitField0_ = (bitField0_ & ~0x00000020);
destOrder_ = 0;
onChanged();
return this;
}
private java.lang.Object ownerName_ = "";
/**
* <code>optional string ownerName = 10;</code>
* @return Whether the ownerName field is set.
*/
public boolean hasOwnerName() {
return ((bitField0_ & 0x00000040) != 0);
}
/**
* <code>optional string ownerName = 10;</code>
* @return The ownerName.
*/
public java.lang.String getOwnerName() {
java.lang.Object ref = ownerName_;
if (!(ref instanceof java.lang.String)) {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
ownerName_ = s;
}
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* <code>optional string ownerName = 10;</code>
* @return The bytes for ownerName.
*/
public org.apache.hadoop.thirdparty.protobuf.ByteString
getOwnerNameBytes() {
java.lang.Object ref = ownerName_;
if (ref instanceof String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
ownerName_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
/**
* <code>optional string ownerName = 10;</code>
* @param value The ownerName to set.
* @return This builder for chaining.
*/
public Builder setOwnerName(
java.lang.String value) {
if (value == null) { throw new NullPointerException(); }
ownerName_ = value;
bitField0_ |= 0x00000040;
onChanged();
return this;
}
/**
* <code>optional string ownerName = 10;</code>
* @return This builder for chaining.
*/
public Builder clearOwnerName() {
ownerName_ = getDefaultInstance().getOwnerName();
bitField0_ = (bitField0_ & ~0x00000040);
onChanged();
return this;
}
/**
* <code>optional string ownerName = 10;</code>
* @param value The bytes for ownerName to set.
* @return This builder for chaining.
*/
public Builder setOwnerNameBytes(
org.apache.hadoop.thirdparty.protobuf.ByteString value) {
if (value == null) { throw new NullPointerException(); }
ownerName_ = value;
bitField0_ |= 0x00000040;
onChanged();
return this;
}
private java.lang.Object groupName_ = "";
/**
* <code>optional string groupName = 11;</code>
* @return Whether the groupName field is set.
*/
public boolean hasGroupName() {
return ((bitField0_ & 0x00000080) != 0);
}
/**
* <code>optional string groupName = 11;</code>
* @return The groupName.
*/
public java.lang.String getGroupName() {
java.lang.Object ref = groupName_;
if (!(ref instanceof java.lang.String)) {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
groupName_ = s;
}
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* <code>optional string groupName = 11;</code>
* @return The bytes for groupName.
*/
public org.apache.hadoop.thirdparty.protobuf.ByteString
getGroupNameBytes() {
java.lang.Object ref = groupName_;
if (ref instanceof String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
groupName_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
/**
* <code>optional string groupName = 11;</code>
* @param value The groupName to set.
* @return This builder for chaining.
*/
public Builder setGroupName(
java.lang.String value) {
if (value == null) { throw new NullPointerException(); }
groupName_ = value;
bitField0_ |= 0x00000080;
onChanged();
return this;
}
/**
* <code>optional string groupName = 11;</code>
* @return This builder for chaining.
*/
public Builder clearGroupName() {
groupName_ = getDefaultInstance().getGroupName();
bitField0_ = (bitField0_ & ~0x00000080);
onChanged();
return this;
}
/**
* <code>optional string groupName = 11;</code>
* @param value The bytes for groupName to set.
* @return This builder for chaining.
*/
public Builder setGroupNameBytes(
org.apache.hadoop.thirdparty.protobuf.ByteString value) {
if (value == null) { throw new NullPointerException(); }
groupName_ = value;
bitField0_ |= 0x00000080;
onChanged();
return this;
}
private int mode_ ;
/**
* <code>optional int32 mode = 12;</code>
* @return Whether the mode field is set.
*/
@java.lang.Override
public boolean hasMode() {
return ((bitField0_ & 0x00000100) != 0);
}
/**
* <code>optional int32 mode = 12;</code>
* @return The mode.
*/
@java.lang.Override
public int getMode() {
return mode_;
}
/**
* <code>optional int32 mode = 12;</code>
* @param value The mode to set.
* @return This builder for chaining.
*/
public Builder setMode(int value) {
mode_ = value;
bitField0_ |= 0x00000100;
onChanged();
return this;
}
/**
* <code>optional int32 mode = 12;</code>
* @return This builder for chaining.
*/
public Builder clearMode() {
bitField0_ = (bitField0_ & ~0x00000100);
mode_ = 0;
onChanged();
return this;
}
private org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.QuotaUsageProto quota_;
private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.QuotaUsageProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.QuotaUsageProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.QuotaUsageProtoOrBuilder> quotaBuilder_;
/**
* <code>optional .hadoop.hdfs.QuotaUsageProto quota = 13;</code>
* @return Whether the quota field is set.
*/
public boolean hasQuota() {
return ((bitField0_ & 0x00000200) != 0);
}
/**
* <code>optional .hadoop.hdfs.QuotaUsageProto quota = 13;</code>
* @return The quota.
*/
public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.QuotaUsageProto getQuota() {
if (quotaBuilder_ == null) {
return quota_ == null ? org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.QuotaUsageProto.getDefaultInstance() : quota_;
} else {
return quotaBuilder_.getMessage();
}
}
/**
* <code>optional .hadoop.hdfs.QuotaUsageProto quota = 13;</code>
*/
public Builder setQuota(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.QuotaUsageProto value) {
if (quotaBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
quota_ = value;
} else {
quotaBuilder_.setMessage(value);
}
bitField0_ |= 0x00000200;
onChanged();
return this;
}
/**
* <code>optional .hadoop.hdfs.QuotaUsageProto quota = 13;</code>
*/
public Builder setQuota(
org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.QuotaUsageProto.Builder builderForValue) {
if (quotaBuilder_ == null) {
quota_ = builderForValue.build();
} else {
quotaBuilder_.setMessage(builderForValue.build());
}
bitField0_ |= 0x00000200;
onChanged();
return this;
}
/**
* <code>optional .hadoop.hdfs.QuotaUsageProto quota = 13;</code>
*/
public Builder mergeQuota(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.QuotaUsageProto value) {
if (quotaBuilder_ == null) {
if (((bitField0_ & 0x00000200) != 0) &&
quota_ != null &&
quota_ != org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.QuotaUsageProto.getDefaultInstance()) {
getQuotaBuilder().mergeFrom(value);
} else {
quota_ = value;
}
} else {
quotaBuilder_.mergeFrom(value);
}
if (quota_ != null) {
bitField0_ |= 0x00000200;
onChanged();
}
return this;
}
/**
* <code>optional .hadoop.hdfs.QuotaUsageProto quota = 13;</code>
*/
public Builder clearQuota() {
bitField0_ = (bitField0_ & ~0x00000200);
quota_ = null;
if (quotaBuilder_ != null) {
quotaBuilder_.dispose();
quotaBuilder_ = null;
}
onChanged();
return this;
}
/**
* <code>optional .hadoop.hdfs.QuotaUsageProto quota = 13;</code>
*/
public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.QuotaUsageProto.Builder getQuotaBuilder() {
bitField0_ |= 0x00000200;
onChanged();
return getQuotaFieldBuilder().getBuilder();
}
/**
* <code>optional .hadoop.hdfs.QuotaUsageProto quota = 13;</code>
*/
public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.QuotaUsageProtoOrBuilder getQuotaOrBuilder() {
if (quotaBuilder_ != null) {
return quotaBuilder_.getMessageOrBuilder();
} else {
return quota_ == null ?
org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.QuotaUsageProto.getDefaultInstance() : quota_;
}
}
/**
* <code>optional .hadoop.hdfs.QuotaUsageProto quota = 13;</code>
*/
private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.QuotaUsageProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.QuotaUsageProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.QuotaUsageProtoOrBuilder>
getQuotaFieldBuilder() {
if (quotaBuilder_ == null) {
quotaBuilder_ = new org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.QuotaUsageProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.QuotaUsageProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.QuotaUsageProtoOrBuilder>(
getQuota(),
getParentForChildren(),
isClean());
quota_ = null;
}
return quotaBuilder_;
}
private boolean faultTolerant_ ;
/**
* <code>optional bool faultTolerant = 14 [default = false];</code>
* @return Whether the faultTolerant field is set.
*/
@java.lang.Override
public boolean hasFaultTolerant() {
return ((bitField0_ & 0x00000400) != 0);
}
/**
* <code>optional bool faultTolerant = 14 [default = false];</code>
* @return The faultTolerant.
*/
@java.lang.Override
public boolean getFaultTolerant() {
return faultTolerant_;
}
/**
* <code>optional bool faultTolerant = 14 [default = false];</code>
* @param value The faultTolerant to set.
* @return This builder for chaining.
*/
public Builder setFaultTolerant(boolean value) {
faultTolerant_ = value;
bitField0_ |= 0x00000400;
onChanged();
return this;
}
/**
* <code>optional bool faultTolerant = 14 [default = false];</code>
* @return This builder for chaining.
*/
public Builder clearFaultTolerant() {
bitField0_ = (bitField0_ & ~0x00000400);
faultTolerant_ = false;
onChanged();
return this;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.MountTableRecordProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.MountTableRecordProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<MountTableRecordProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<MountTableRecordProto>() {
@java.lang.Override
public MountTableRecordProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<MountTableRecordProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<MountTableRecordProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface AddMountTableEntryRequestProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.AddMountTableEntryRequestProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* <code>optional .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
* @return Whether the entry field is set.
*/
boolean hasEntry();
/**
* <code>optional .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
* @return The entry.
*/
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto getEntry();
/**
* <code>optional .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
*/
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProtoOrBuilder getEntryOrBuilder();
}
/**
* Protobuf type {@code hadoop.hdfs.AddMountTableEntryRequestProto}
*/
public static final class AddMountTableEntryRequestProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.AddMountTableEntryRequestProto)
AddMountTableEntryRequestProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use AddMountTableEntryRequestProto.newBuilder() to construct.
private AddMountTableEntryRequestProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private AddMountTableEntryRequestProto() {
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new AddMountTableEntryRequestProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_AddMountTableEntryRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_AddMountTableEntryRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryRequestProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryRequestProto.Builder.class);
}
private int bitField0_;
public static final int ENTRY_FIELD_NUMBER = 1;
private org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto entry_;
/**
* <code>optional .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
* @return Whether the entry field is set.
*/
@java.lang.Override
public boolean hasEntry() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
* @return The entry.
*/
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto getEntry() {
return entry_ == null ? org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.getDefaultInstance() : entry_;
}
/**
* <code>optional .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
*/
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProtoOrBuilder getEntryOrBuilder() {
return entry_ == null ? org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.getDefaultInstance() : entry_;
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
if (hasEntry()) {
if (!getEntry().isInitialized()) {
memoizedIsInitialized = 0;
return false;
}
}
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) != 0)) {
output.writeMessage(1, getEntry());
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeMessageSize(1, getEntry());
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryRequestProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryRequestProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryRequestProto) obj;
if (hasEntry() != other.hasEntry()) return false;
if (hasEntry()) {
if (!getEntry()
.equals(other.getEntry())) return false;
}
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (hasEntry()) {
hash = (37 * hash) + ENTRY_FIELD_NUMBER;
hash = (53 * hash) + getEntry().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryRequestProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryRequestProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryRequestProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryRequestProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryRequestProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryRequestProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryRequestProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryRequestProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryRequestProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.AddMountTableEntryRequestProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.AddMountTableEntryRequestProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryRequestProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_AddMountTableEntryRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_AddMountTableEntryRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryRequestProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryRequestProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryRequestProto.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
if (org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
getEntryFieldBuilder();
}
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
entry_ = null;
if (entryBuilder_ != null) {
entryBuilder_.dispose();
entryBuilder_ = null;
}
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_AddMountTableEntryRequestProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryRequestProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryRequestProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryRequestProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryRequestProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryRequestProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryRequestProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryRequestProto(this);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartial0(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryRequestProto result) {
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) != 0)) {
result.entry_ = entryBuilder_ == null
? entry_
: entryBuilder_.build();
to_bitField0_ |= 0x00000001;
}
result.bitField0_ |= to_bitField0_;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryRequestProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryRequestProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryRequestProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryRequestProto.getDefaultInstance()) return this;
if (other.hasEntry()) {
mergeEntry(other.getEntry());
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
if (hasEntry()) {
if (!getEntry().isInitialized()) {
return false;
}
}
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 10: {
input.readMessage(
getEntryFieldBuilder().getBuilder(),
extensionRegistry);
bitField0_ |= 0x00000001;
break;
} // case 10
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto entry_;
private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.Builder, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProtoOrBuilder> entryBuilder_;
/**
* <code>optional .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
* @return Whether the entry field is set.
*/
public boolean hasEntry() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
* @return The entry.
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto getEntry() {
if (entryBuilder_ == null) {
return entry_ == null ? org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.getDefaultInstance() : entry_;
} else {
return entryBuilder_.getMessage();
}
}
/**
* <code>optional .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
*/
public Builder setEntry(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto value) {
if (entryBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
entry_ = value;
} else {
entryBuilder_.setMessage(value);
}
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>optional .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
*/
public Builder setEntry(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.Builder builderForValue) {
if (entryBuilder_ == null) {
entry_ = builderForValue.build();
} else {
entryBuilder_.setMessage(builderForValue.build());
}
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>optional .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
*/
public Builder mergeEntry(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto value) {
if (entryBuilder_ == null) {
if (((bitField0_ & 0x00000001) != 0) &&
entry_ != null &&
entry_ != org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.getDefaultInstance()) {
getEntryBuilder().mergeFrom(value);
} else {
entry_ = value;
}
} else {
entryBuilder_.mergeFrom(value);
}
if (entry_ != null) {
bitField0_ |= 0x00000001;
onChanged();
}
return this;
}
/**
* <code>optional .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
*/
public Builder clearEntry() {
bitField0_ = (bitField0_ & ~0x00000001);
entry_ = null;
if (entryBuilder_ != null) {
entryBuilder_.dispose();
entryBuilder_ = null;
}
onChanged();
return this;
}
/**
* <code>optional .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.Builder getEntryBuilder() {
bitField0_ |= 0x00000001;
onChanged();
return getEntryFieldBuilder().getBuilder();
}
/**
* <code>optional .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProtoOrBuilder getEntryOrBuilder() {
if (entryBuilder_ != null) {
return entryBuilder_.getMessageOrBuilder();
} else {
return entry_ == null ?
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.getDefaultInstance() : entry_;
}
}
/**
* <code>optional .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
*/
private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.Builder, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProtoOrBuilder>
getEntryFieldBuilder() {
if (entryBuilder_ == null) {
entryBuilder_ = new org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.Builder, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProtoOrBuilder>(
getEntry(),
getParentForChildren(),
isClean());
entry_ = null;
}
return entryBuilder_;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.AddMountTableEntryRequestProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.AddMountTableEntryRequestProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryRequestProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryRequestProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryRequestProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<AddMountTableEntryRequestProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<AddMountTableEntryRequestProto>() {
@java.lang.Override
public AddMountTableEntryRequestProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<AddMountTableEntryRequestProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<AddMountTableEntryRequestProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryRequestProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface AddMountTableEntryResponseProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.AddMountTableEntryResponseProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* <code>optional bool status = 1;</code>
* @return Whether the status field is set.
*/
boolean hasStatus();
/**
* <code>optional bool status = 1;</code>
* @return The status.
*/
boolean getStatus();
}
/**
* Protobuf type {@code hadoop.hdfs.AddMountTableEntryResponseProto}
*/
public static final class AddMountTableEntryResponseProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.AddMountTableEntryResponseProto)
AddMountTableEntryResponseProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use AddMountTableEntryResponseProto.newBuilder() to construct.
private AddMountTableEntryResponseProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private AddMountTableEntryResponseProto() {
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new AddMountTableEntryResponseProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_AddMountTableEntryResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_AddMountTableEntryResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryResponseProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryResponseProto.Builder.class);
}
private int bitField0_;
public static final int STATUS_FIELD_NUMBER = 1;
private boolean status_ = false;
/**
* <code>optional bool status = 1;</code>
* @return Whether the status field is set.
*/
@java.lang.Override
public boolean hasStatus() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional bool status = 1;</code>
* @return The status.
*/
@java.lang.Override
public boolean getStatus() {
return status_;
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) != 0)) {
output.writeBool(1, status_);
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeBoolSize(1, status_);
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryResponseProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryResponseProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryResponseProto) obj;
if (hasStatus() != other.hasStatus()) return false;
if (hasStatus()) {
if (getStatus()
!= other.getStatus()) return false;
}
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (hasStatus()) {
hash = (37 * hash) + STATUS_FIELD_NUMBER;
hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashBoolean(
getStatus());
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryResponseProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryResponseProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryResponseProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryResponseProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryResponseProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryResponseProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryResponseProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryResponseProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryResponseProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.AddMountTableEntryResponseProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.AddMountTableEntryResponseProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryResponseProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_AddMountTableEntryResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_AddMountTableEntryResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryResponseProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryResponseProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryResponseProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
status_ = false;
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_AddMountTableEntryResponseProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryResponseProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryResponseProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryResponseProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryResponseProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryResponseProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryResponseProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryResponseProto(this);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartial0(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryResponseProto result) {
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) != 0)) {
result.status_ = status_;
to_bitField0_ |= 0x00000001;
}
result.bitField0_ |= to_bitField0_;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryResponseProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryResponseProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryResponseProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryResponseProto.getDefaultInstance()) return this;
if (other.hasStatus()) {
setStatus(other.getStatus());
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 8: {
status_ = input.readBool();
bitField0_ |= 0x00000001;
break;
} // case 8
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private boolean status_ ;
/**
* <code>optional bool status = 1;</code>
* @return Whether the status field is set.
*/
@java.lang.Override
public boolean hasStatus() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional bool status = 1;</code>
* @return The status.
*/
@java.lang.Override
public boolean getStatus() {
return status_;
}
/**
* <code>optional bool status = 1;</code>
* @param value The status to set.
* @return This builder for chaining.
*/
public Builder setStatus(boolean value) {
status_ = value;
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>optional bool status = 1;</code>
* @return This builder for chaining.
*/
public Builder clearStatus() {
bitField0_ = (bitField0_ & ~0x00000001);
status_ = false;
onChanged();
return this;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.AddMountTableEntryResponseProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.AddMountTableEntryResponseProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryResponseProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryResponseProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryResponseProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<AddMountTableEntryResponseProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<AddMountTableEntryResponseProto>() {
@java.lang.Override
public AddMountTableEntryResponseProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<AddMountTableEntryResponseProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<AddMountTableEntryResponseProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntryResponseProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface AddMountTableEntriesRequestProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.AddMountTableEntriesRequestProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
*/
java.util.List<org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto>
getEntryList();
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
*/
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto getEntry(int index);
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
*/
int getEntryCount();
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
*/
java.util.List<? extends org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProtoOrBuilder>
getEntryOrBuilderList();
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
*/
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProtoOrBuilder getEntryOrBuilder(
int index);
}
/**
* Protobuf type {@code hadoop.hdfs.AddMountTableEntriesRequestProto}
*/
public static final class AddMountTableEntriesRequestProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.AddMountTableEntriesRequestProto)
AddMountTableEntriesRequestProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use AddMountTableEntriesRequestProto.newBuilder() to construct.
private AddMountTableEntriesRequestProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private AddMountTableEntriesRequestProto() {
entry_ = java.util.Collections.emptyList();
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new AddMountTableEntriesRequestProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_AddMountTableEntriesRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_AddMountTableEntriesRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesRequestProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesRequestProto.Builder.class);
}
public static final int ENTRY_FIELD_NUMBER = 1;
@SuppressWarnings("serial")
private java.util.List<org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto> entry_;
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
*/
@java.lang.Override
public java.util.List<org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto> getEntryList() {
return entry_;
}
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
*/
@java.lang.Override
public java.util.List<? extends org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProtoOrBuilder>
getEntryOrBuilderList() {
return entry_;
}
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
*/
@java.lang.Override
public int getEntryCount() {
return entry_.size();
}
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
*/
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto getEntry(int index) {
return entry_.get(index);
}
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
*/
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProtoOrBuilder getEntryOrBuilder(
int index) {
return entry_.get(index);
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
for (int i = 0; i < getEntryCount(); i++) {
if (!getEntry(i).isInitialized()) {
memoizedIsInitialized = 0;
return false;
}
}
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
for (int i = 0; i < entry_.size(); i++) {
output.writeMessage(1, entry_.get(i));
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
for (int i = 0; i < entry_.size(); i++) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeMessageSize(1, entry_.get(i));
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesRequestProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesRequestProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesRequestProto) obj;
if (!getEntryList()
.equals(other.getEntryList())) return false;
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (getEntryCount() > 0) {
hash = (37 * hash) + ENTRY_FIELD_NUMBER;
hash = (53 * hash) + getEntryList().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesRequestProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesRequestProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesRequestProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesRequestProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesRequestProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesRequestProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesRequestProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesRequestProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesRequestProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.AddMountTableEntriesRequestProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.AddMountTableEntriesRequestProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesRequestProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_AddMountTableEntriesRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_AddMountTableEntriesRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesRequestProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesRequestProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesRequestProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
if (entryBuilder_ == null) {
entry_ = java.util.Collections.emptyList();
} else {
entry_ = null;
entryBuilder_.clear();
}
bitField0_ = (bitField0_ & ~0x00000001);
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_AddMountTableEntriesRequestProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesRequestProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesRequestProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesRequestProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesRequestProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesRequestProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesRequestProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesRequestProto(this);
buildPartialRepeatedFields(result);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartialRepeatedFields(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesRequestProto result) {
if (entryBuilder_ == null) {
if (((bitField0_ & 0x00000001) != 0)) {
entry_ = java.util.Collections.unmodifiableList(entry_);
bitField0_ = (bitField0_ & ~0x00000001);
}
result.entry_ = entry_;
} else {
result.entry_ = entryBuilder_.build();
}
}
private void buildPartial0(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesRequestProto result) {
int from_bitField0_ = bitField0_;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesRequestProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesRequestProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesRequestProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesRequestProto.getDefaultInstance()) return this;
if (entryBuilder_ == null) {
if (!other.entry_.isEmpty()) {
if (entry_.isEmpty()) {
entry_ = other.entry_;
bitField0_ = (bitField0_ & ~0x00000001);
} else {
ensureEntryIsMutable();
entry_.addAll(other.entry_);
}
onChanged();
}
} else {
if (!other.entry_.isEmpty()) {
if (entryBuilder_.isEmpty()) {
entryBuilder_.dispose();
entryBuilder_ = null;
entry_ = other.entry_;
bitField0_ = (bitField0_ & ~0x00000001);
entryBuilder_ =
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
getEntryFieldBuilder() : null;
} else {
entryBuilder_.addAllMessages(other.entry_);
}
}
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
for (int i = 0; i < getEntryCount(); i++) {
if (!getEntry(i).isInitialized()) {
return false;
}
}
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 10: {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto m =
input.readMessage(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.PARSER,
extensionRegistry);
if (entryBuilder_ == null) {
ensureEntryIsMutable();
entry_.add(m);
} else {
entryBuilder_.addMessage(m);
}
break;
} // case 10
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private java.util.List<org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto> entry_ =
java.util.Collections.emptyList();
private void ensureEntryIsMutable() {
if (!((bitField0_ & 0x00000001) != 0)) {
entry_ = new java.util.ArrayList<org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto>(entry_);
bitField0_ |= 0x00000001;
}
}
private org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.Builder, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProtoOrBuilder> entryBuilder_;
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
*/
public java.util.List<org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto> getEntryList() {
if (entryBuilder_ == null) {
return java.util.Collections.unmodifiableList(entry_);
} else {
return entryBuilder_.getMessageList();
}
}
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
*/
public int getEntryCount() {
if (entryBuilder_ == null) {
return entry_.size();
} else {
return entryBuilder_.getCount();
}
}
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto getEntry(int index) {
if (entryBuilder_ == null) {
return entry_.get(index);
} else {
return entryBuilder_.getMessage(index);
}
}
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
*/
public Builder setEntry(
int index, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto value) {
if (entryBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
ensureEntryIsMutable();
entry_.set(index, value);
onChanged();
} else {
entryBuilder_.setMessage(index, value);
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
*/
public Builder setEntry(
int index, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.Builder builderForValue) {
if (entryBuilder_ == null) {
ensureEntryIsMutable();
entry_.set(index, builderForValue.build());
onChanged();
} else {
entryBuilder_.setMessage(index, builderForValue.build());
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
*/
public Builder addEntry(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto value) {
if (entryBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
ensureEntryIsMutable();
entry_.add(value);
onChanged();
} else {
entryBuilder_.addMessage(value);
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
*/
public Builder addEntry(
int index, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto value) {
if (entryBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
ensureEntryIsMutable();
entry_.add(index, value);
onChanged();
} else {
entryBuilder_.addMessage(index, value);
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
*/
public Builder addEntry(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.Builder builderForValue) {
if (entryBuilder_ == null) {
ensureEntryIsMutable();
entry_.add(builderForValue.build());
onChanged();
} else {
entryBuilder_.addMessage(builderForValue.build());
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
*/
public Builder addEntry(
int index, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.Builder builderForValue) {
if (entryBuilder_ == null) {
ensureEntryIsMutable();
entry_.add(index, builderForValue.build());
onChanged();
} else {
entryBuilder_.addMessage(index, builderForValue.build());
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
*/
public Builder addAllEntry(
java.lang.Iterable<? extends org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto> values) {
if (entryBuilder_ == null) {
ensureEntryIsMutable();
org.apache.hadoop.thirdparty.protobuf.AbstractMessageLite.Builder.addAll(
values, entry_);
onChanged();
} else {
entryBuilder_.addAllMessages(values);
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
*/
public Builder clearEntry() {
if (entryBuilder_ == null) {
entry_ = java.util.Collections.emptyList();
bitField0_ = (bitField0_ & ~0x00000001);
onChanged();
} else {
entryBuilder_.clear();
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
*/
public Builder removeEntry(int index) {
if (entryBuilder_ == null) {
ensureEntryIsMutable();
entry_.remove(index);
onChanged();
} else {
entryBuilder_.remove(index);
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.Builder getEntryBuilder(
int index) {
return getEntryFieldBuilder().getBuilder(index);
}
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProtoOrBuilder getEntryOrBuilder(
int index) {
if (entryBuilder_ == null) {
return entry_.get(index); } else {
return entryBuilder_.getMessageOrBuilder(index);
}
}
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
*/
public java.util.List<? extends org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProtoOrBuilder>
getEntryOrBuilderList() {
if (entryBuilder_ != null) {
return entryBuilder_.getMessageOrBuilderList();
} else {
return java.util.Collections.unmodifiableList(entry_);
}
}
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.Builder addEntryBuilder() {
return getEntryFieldBuilder().addBuilder(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.getDefaultInstance());
}
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.Builder addEntryBuilder(
int index) {
return getEntryFieldBuilder().addBuilder(
index, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.getDefaultInstance());
}
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
*/
public java.util.List<org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.Builder>
getEntryBuilderList() {
return getEntryFieldBuilder().getBuilderList();
}
private org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.Builder, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProtoOrBuilder>
getEntryFieldBuilder() {
if (entryBuilder_ == null) {
entryBuilder_ = new org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.Builder, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProtoOrBuilder>(
entry_,
((bitField0_ & 0x00000001) != 0),
getParentForChildren(),
isClean());
entry_ = null;
}
return entryBuilder_;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.AddMountTableEntriesRequestProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.AddMountTableEntriesRequestProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesRequestProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesRequestProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesRequestProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<AddMountTableEntriesRequestProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<AddMountTableEntriesRequestProto>() {
@java.lang.Override
public AddMountTableEntriesRequestProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<AddMountTableEntriesRequestProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<AddMountTableEntriesRequestProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesRequestProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface AddMountTableEntriesResponseProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.AddMountTableEntriesResponseProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* <code>optional bool status = 1;</code>
* @return Whether the status field is set.
*/
boolean hasStatus();
/**
* <code>optional bool status = 1;</code>
* @return The status.
*/
boolean getStatus();
/**
* <code>repeated string failedEntriesKeys = 2;</code>
* @return A list containing the failedEntriesKeys.
*/
java.util.List<java.lang.String>
getFailedEntriesKeysList();
/**
* <code>repeated string failedEntriesKeys = 2;</code>
* @return The count of failedEntriesKeys.
*/
int getFailedEntriesKeysCount();
/**
* <code>repeated string failedEntriesKeys = 2;</code>
* @param index The index of the element to return.
* @return The failedEntriesKeys at the given index.
*/
java.lang.String getFailedEntriesKeys(int index);
/**
* <code>repeated string failedEntriesKeys = 2;</code>
* @param index The index of the value to return.
* @return The bytes of the failedEntriesKeys at the given index.
*/
org.apache.hadoop.thirdparty.protobuf.ByteString
getFailedEntriesKeysBytes(int index);
}
/**
* Protobuf type {@code hadoop.hdfs.AddMountTableEntriesResponseProto}
*/
public static final class AddMountTableEntriesResponseProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.AddMountTableEntriesResponseProto)
AddMountTableEntriesResponseProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use AddMountTableEntriesResponseProto.newBuilder() to construct.
private AddMountTableEntriesResponseProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private AddMountTableEntriesResponseProto() {
failedEntriesKeys_ =
org.apache.hadoop.thirdparty.protobuf.LazyStringArrayList.emptyList();
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new AddMountTableEntriesResponseProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_AddMountTableEntriesResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_AddMountTableEntriesResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesResponseProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesResponseProto.Builder.class);
}
private int bitField0_;
public static final int STATUS_FIELD_NUMBER = 1;
private boolean status_ = false;
/**
* <code>optional bool status = 1;</code>
* @return Whether the status field is set.
*/
@java.lang.Override
public boolean hasStatus() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional bool status = 1;</code>
* @return The status.
*/
@java.lang.Override
public boolean getStatus() {
return status_;
}
public static final int FAILEDENTRIESKEYS_FIELD_NUMBER = 2;
@SuppressWarnings("serial")
private org.apache.hadoop.thirdparty.protobuf.LazyStringArrayList failedEntriesKeys_ =
org.apache.hadoop.thirdparty.protobuf.LazyStringArrayList.emptyList();
/**
* <code>repeated string failedEntriesKeys = 2;</code>
* @return A list containing the failedEntriesKeys.
*/
public org.apache.hadoop.thirdparty.protobuf.ProtocolStringList
getFailedEntriesKeysList() {
return failedEntriesKeys_;
}
/**
* <code>repeated string failedEntriesKeys = 2;</code>
* @return The count of failedEntriesKeys.
*/
public int getFailedEntriesKeysCount() {
return failedEntriesKeys_.size();
}
/**
* <code>repeated string failedEntriesKeys = 2;</code>
* @param index The index of the element to return.
* @return The failedEntriesKeys at the given index.
*/
public java.lang.String getFailedEntriesKeys(int index) {
return failedEntriesKeys_.get(index);
}
/**
* <code>repeated string failedEntriesKeys = 2;</code>
* @param index The index of the value to return.
* @return The bytes of the failedEntriesKeys at the given index.
*/
public org.apache.hadoop.thirdparty.protobuf.ByteString
getFailedEntriesKeysBytes(int index) {
return failedEntriesKeys_.getByteString(index);
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) != 0)) {
output.writeBool(1, status_);
}
for (int i = 0; i < failedEntriesKeys_.size(); i++) {
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 2, failedEntriesKeys_.getRaw(i));
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeBoolSize(1, status_);
}
{
int dataSize = 0;
for (int i = 0; i < failedEntriesKeys_.size(); i++) {
dataSize += computeStringSizeNoTag(failedEntriesKeys_.getRaw(i));
}
size += dataSize;
size += 1 * getFailedEntriesKeysList().size();
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesResponseProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesResponseProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesResponseProto) obj;
if (hasStatus() != other.hasStatus()) return false;
if (hasStatus()) {
if (getStatus()
!= other.getStatus()) return false;
}
if (!getFailedEntriesKeysList()
.equals(other.getFailedEntriesKeysList())) return false;
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (hasStatus()) {
hash = (37 * hash) + STATUS_FIELD_NUMBER;
hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashBoolean(
getStatus());
}
if (getFailedEntriesKeysCount() > 0) {
hash = (37 * hash) + FAILEDENTRIESKEYS_FIELD_NUMBER;
hash = (53 * hash) + getFailedEntriesKeysList().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesResponseProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesResponseProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesResponseProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesResponseProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesResponseProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesResponseProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesResponseProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesResponseProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesResponseProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.AddMountTableEntriesResponseProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.AddMountTableEntriesResponseProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesResponseProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_AddMountTableEntriesResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_AddMountTableEntriesResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesResponseProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesResponseProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesResponseProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
status_ = false;
failedEntriesKeys_ =
org.apache.hadoop.thirdparty.protobuf.LazyStringArrayList.emptyList();
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_AddMountTableEntriesResponseProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesResponseProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesResponseProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesResponseProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesResponseProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesResponseProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesResponseProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesResponseProto(this);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartial0(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesResponseProto result) {
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) != 0)) {
result.status_ = status_;
to_bitField0_ |= 0x00000001;
}
if (((from_bitField0_ & 0x00000002) != 0)) {
failedEntriesKeys_.makeImmutable();
result.failedEntriesKeys_ = failedEntriesKeys_;
}
result.bitField0_ |= to_bitField0_;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesResponseProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesResponseProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesResponseProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesResponseProto.getDefaultInstance()) return this;
if (other.hasStatus()) {
setStatus(other.getStatus());
}
if (!other.failedEntriesKeys_.isEmpty()) {
if (failedEntriesKeys_.isEmpty()) {
failedEntriesKeys_ = other.failedEntriesKeys_;
bitField0_ |= 0x00000002;
} else {
ensureFailedEntriesKeysIsMutable();
failedEntriesKeys_.addAll(other.failedEntriesKeys_);
}
onChanged();
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 8: {
status_ = input.readBool();
bitField0_ |= 0x00000001;
break;
} // case 8
case 18: {
org.apache.hadoop.thirdparty.protobuf.ByteString bs = input.readBytes();
ensureFailedEntriesKeysIsMutable();
failedEntriesKeys_.add(bs);
break;
} // case 18
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private boolean status_ ;
/**
* <code>optional bool status = 1;</code>
* @return Whether the status field is set.
*/
@java.lang.Override
public boolean hasStatus() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional bool status = 1;</code>
* @return The status.
*/
@java.lang.Override
public boolean getStatus() {
return status_;
}
/**
* <code>optional bool status = 1;</code>
* @param value The status to set.
* @return This builder for chaining.
*/
public Builder setStatus(boolean value) {
status_ = value;
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>optional bool status = 1;</code>
* @return This builder for chaining.
*/
public Builder clearStatus() {
bitField0_ = (bitField0_ & ~0x00000001);
status_ = false;
onChanged();
return this;
}
private org.apache.hadoop.thirdparty.protobuf.LazyStringArrayList failedEntriesKeys_ =
org.apache.hadoop.thirdparty.protobuf.LazyStringArrayList.emptyList();
private void ensureFailedEntriesKeysIsMutable() {
if (!failedEntriesKeys_.isModifiable()) {
failedEntriesKeys_ = new org.apache.hadoop.thirdparty.protobuf.LazyStringArrayList(failedEntriesKeys_);
}
bitField0_ |= 0x00000002;
}
/**
* <code>repeated string failedEntriesKeys = 2;</code>
* @return A list containing the failedEntriesKeys.
*/
public org.apache.hadoop.thirdparty.protobuf.ProtocolStringList
getFailedEntriesKeysList() {
failedEntriesKeys_.makeImmutable();
return failedEntriesKeys_;
}
/**
* <code>repeated string failedEntriesKeys = 2;</code>
* @return The count of failedEntriesKeys.
*/
public int getFailedEntriesKeysCount() {
return failedEntriesKeys_.size();
}
/**
* <code>repeated string failedEntriesKeys = 2;</code>
* @param index The index of the element to return.
* @return The failedEntriesKeys at the given index.
*/
public java.lang.String getFailedEntriesKeys(int index) {
return failedEntriesKeys_.get(index);
}
/**
* <code>repeated string failedEntriesKeys = 2;</code>
* @param index The index of the value to return.
* @return The bytes of the failedEntriesKeys at the given index.
*/
public org.apache.hadoop.thirdparty.protobuf.ByteString
getFailedEntriesKeysBytes(int index) {
return failedEntriesKeys_.getByteString(index);
}
/**
* <code>repeated string failedEntriesKeys = 2;</code>
* @param index The index to set the value at.
* @param value The failedEntriesKeys to set.
* @return This builder for chaining.
*/
public Builder setFailedEntriesKeys(
int index, java.lang.String value) {
if (value == null) { throw new NullPointerException(); }
ensureFailedEntriesKeysIsMutable();
failedEntriesKeys_.set(index, value);
bitField0_ |= 0x00000002;
onChanged();
return this;
}
/**
* <code>repeated string failedEntriesKeys = 2;</code>
* @param value The failedEntriesKeys to add.
* @return This builder for chaining.
*/
public Builder addFailedEntriesKeys(
java.lang.String value) {
if (value == null) { throw new NullPointerException(); }
ensureFailedEntriesKeysIsMutable();
failedEntriesKeys_.add(value);
bitField0_ |= 0x00000002;
onChanged();
return this;
}
/**
* <code>repeated string failedEntriesKeys = 2;</code>
* @param values The failedEntriesKeys to add.
* @return This builder for chaining.
*/
public Builder addAllFailedEntriesKeys(
java.lang.Iterable<java.lang.String> values) {
ensureFailedEntriesKeysIsMutable();
org.apache.hadoop.thirdparty.protobuf.AbstractMessageLite.Builder.addAll(
values, failedEntriesKeys_);
bitField0_ |= 0x00000002;
onChanged();
return this;
}
/**
* <code>repeated string failedEntriesKeys = 2;</code>
* @return This builder for chaining.
*/
public Builder clearFailedEntriesKeys() {
failedEntriesKeys_ =
org.apache.hadoop.thirdparty.protobuf.LazyStringArrayList.emptyList();
bitField0_ = (bitField0_ & ~0x00000002);;
onChanged();
return this;
}
/**
* <code>repeated string failedEntriesKeys = 2;</code>
* @param value The bytes of the failedEntriesKeys to add.
* @return This builder for chaining.
*/
public Builder addFailedEntriesKeysBytes(
org.apache.hadoop.thirdparty.protobuf.ByteString value) {
if (value == null) { throw new NullPointerException(); }
ensureFailedEntriesKeysIsMutable();
failedEntriesKeys_.add(value);
bitField0_ |= 0x00000002;
onChanged();
return this;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.AddMountTableEntriesResponseProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.AddMountTableEntriesResponseProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesResponseProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesResponseProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesResponseProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<AddMountTableEntriesResponseProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<AddMountTableEntriesResponseProto>() {
@java.lang.Override
public AddMountTableEntriesResponseProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<AddMountTableEntriesResponseProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<AddMountTableEntriesResponseProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.AddMountTableEntriesResponseProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface UpdateMountTableEntryRequestProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.UpdateMountTableEntryRequestProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* <code>optional .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
* @return Whether the entry field is set.
*/
boolean hasEntry();
/**
* <code>optional .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
* @return The entry.
*/
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto getEntry();
/**
* <code>optional .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
*/
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProtoOrBuilder getEntryOrBuilder();
}
/**
* Protobuf type {@code hadoop.hdfs.UpdateMountTableEntryRequestProto}
*/
public static final class UpdateMountTableEntryRequestProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.UpdateMountTableEntryRequestProto)
UpdateMountTableEntryRequestProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use UpdateMountTableEntryRequestProto.newBuilder() to construct.
private UpdateMountTableEntryRequestProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private UpdateMountTableEntryRequestProto() {
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new UpdateMountTableEntryRequestProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_UpdateMountTableEntryRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_UpdateMountTableEntryRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryRequestProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryRequestProto.Builder.class);
}
private int bitField0_;
public static final int ENTRY_FIELD_NUMBER = 1;
private org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto entry_;
/**
* <code>optional .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
* @return Whether the entry field is set.
*/
@java.lang.Override
public boolean hasEntry() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
* @return The entry.
*/
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto getEntry() {
return entry_ == null ? org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.getDefaultInstance() : entry_;
}
/**
* <code>optional .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
*/
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProtoOrBuilder getEntryOrBuilder() {
return entry_ == null ? org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.getDefaultInstance() : entry_;
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
if (hasEntry()) {
if (!getEntry().isInitialized()) {
memoizedIsInitialized = 0;
return false;
}
}
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) != 0)) {
output.writeMessage(1, getEntry());
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeMessageSize(1, getEntry());
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryRequestProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryRequestProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryRequestProto) obj;
if (hasEntry() != other.hasEntry()) return false;
if (hasEntry()) {
if (!getEntry()
.equals(other.getEntry())) return false;
}
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (hasEntry()) {
hash = (37 * hash) + ENTRY_FIELD_NUMBER;
hash = (53 * hash) + getEntry().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryRequestProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryRequestProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryRequestProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryRequestProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryRequestProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryRequestProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryRequestProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryRequestProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryRequestProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.UpdateMountTableEntryRequestProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.UpdateMountTableEntryRequestProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryRequestProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_UpdateMountTableEntryRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_UpdateMountTableEntryRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryRequestProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryRequestProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryRequestProto.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
if (org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
getEntryFieldBuilder();
}
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
entry_ = null;
if (entryBuilder_ != null) {
entryBuilder_.dispose();
entryBuilder_ = null;
}
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_UpdateMountTableEntryRequestProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryRequestProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryRequestProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryRequestProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryRequestProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryRequestProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryRequestProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryRequestProto(this);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartial0(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryRequestProto result) {
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) != 0)) {
result.entry_ = entryBuilder_ == null
? entry_
: entryBuilder_.build();
to_bitField0_ |= 0x00000001;
}
result.bitField0_ |= to_bitField0_;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryRequestProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryRequestProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryRequestProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryRequestProto.getDefaultInstance()) return this;
if (other.hasEntry()) {
mergeEntry(other.getEntry());
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
if (hasEntry()) {
if (!getEntry().isInitialized()) {
return false;
}
}
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 10: {
input.readMessage(
getEntryFieldBuilder().getBuilder(),
extensionRegistry);
bitField0_ |= 0x00000001;
break;
} // case 10
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto entry_;
private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.Builder, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProtoOrBuilder> entryBuilder_;
/**
* <code>optional .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
* @return Whether the entry field is set.
*/
public boolean hasEntry() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
* @return The entry.
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto getEntry() {
if (entryBuilder_ == null) {
return entry_ == null ? org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.getDefaultInstance() : entry_;
} else {
return entryBuilder_.getMessage();
}
}
/**
* <code>optional .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
*/
public Builder setEntry(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto value) {
if (entryBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
entry_ = value;
} else {
entryBuilder_.setMessage(value);
}
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>optional .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
*/
public Builder setEntry(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.Builder builderForValue) {
if (entryBuilder_ == null) {
entry_ = builderForValue.build();
} else {
entryBuilder_.setMessage(builderForValue.build());
}
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>optional .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
*/
public Builder mergeEntry(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto value) {
if (entryBuilder_ == null) {
if (((bitField0_ & 0x00000001) != 0) &&
entry_ != null &&
entry_ != org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.getDefaultInstance()) {
getEntryBuilder().mergeFrom(value);
} else {
entry_ = value;
}
} else {
entryBuilder_.mergeFrom(value);
}
if (entry_ != null) {
bitField0_ |= 0x00000001;
onChanged();
}
return this;
}
/**
* <code>optional .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
*/
public Builder clearEntry() {
bitField0_ = (bitField0_ & ~0x00000001);
entry_ = null;
if (entryBuilder_ != null) {
entryBuilder_.dispose();
entryBuilder_ = null;
}
onChanged();
return this;
}
/**
* <code>optional .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.Builder getEntryBuilder() {
bitField0_ |= 0x00000001;
onChanged();
return getEntryFieldBuilder().getBuilder();
}
/**
* <code>optional .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProtoOrBuilder getEntryOrBuilder() {
if (entryBuilder_ != null) {
return entryBuilder_.getMessageOrBuilder();
} else {
return entry_ == null ?
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.getDefaultInstance() : entry_;
}
}
/**
* <code>optional .hadoop.hdfs.MountTableRecordProto entry = 1;</code>
*/
private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.Builder, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProtoOrBuilder>
getEntryFieldBuilder() {
if (entryBuilder_ == null) {
entryBuilder_ = new org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.Builder, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProtoOrBuilder>(
getEntry(),
getParentForChildren(),
isClean());
entry_ = null;
}
return entryBuilder_;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.UpdateMountTableEntryRequestProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.UpdateMountTableEntryRequestProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryRequestProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryRequestProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryRequestProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<UpdateMountTableEntryRequestProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<UpdateMountTableEntryRequestProto>() {
@java.lang.Override
public UpdateMountTableEntryRequestProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<UpdateMountTableEntryRequestProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<UpdateMountTableEntryRequestProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryRequestProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface UpdateMountTableEntryResponseProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.UpdateMountTableEntryResponseProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* <code>optional bool status = 1;</code>
* @return Whether the status field is set.
*/
boolean hasStatus();
/**
* <code>optional bool status = 1;</code>
* @return The status.
*/
boolean getStatus();
}
/**
* Protobuf type {@code hadoop.hdfs.UpdateMountTableEntryResponseProto}
*/
public static final class UpdateMountTableEntryResponseProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.UpdateMountTableEntryResponseProto)
UpdateMountTableEntryResponseProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use UpdateMountTableEntryResponseProto.newBuilder() to construct.
private UpdateMountTableEntryResponseProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private UpdateMountTableEntryResponseProto() {
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new UpdateMountTableEntryResponseProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_UpdateMountTableEntryResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_UpdateMountTableEntryResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryResponseProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryResponseProto.Builder.class);
}
private int bitField0_;
public static final int STATUS_FIELD_NUMBER = 1;
private boolean status_ = false;
/**
* <code>optional bool status = 1;</code>
* @return Whether the status field is set.
*/
@java.lang.Override
public boolean hasStatus() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional bool status = 1;</code>
* @return The status.
*/
@java.lang.Override
public boolean getStatus() {
return status_;
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) != 0)) {
output.writeBool(1, status_);
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeBoolSize(1, status_);
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryResponseProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryResponseProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryResponseProto) obj;
if (hasStatus() != other.hasStatus()) return false;
if (hasStatus()) {
if (getStatus()
!= other.getStatus()) return false;
}
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (hasStatus()) {
hash = (37 * hash) + STATUS_FIELD_NUMBER;
hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashBoolean(
getStatus());
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryResponseProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryResponseProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryResponseProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryResponseProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryResponseProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryResponseProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryResponseProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryResponseProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryResponseProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.UpdateMountTableEntryResponseProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.UpdateMountTableEntryResponseProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryResponseProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_UpdateMountTableEntryResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_UpdateMountTableEntryResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryResponseProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryResponseProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryResponseProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
status_ = false;
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_UpdateMountTableEntryResponseProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryResponseProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryResponseProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryResponseProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryResponseProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryResponseProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryResponseProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryResponseProto(this);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartial0(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryResponseProto result) {
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) != 0)) {
result.status_ = status_;
to_bitField0_ |= 0x00000001;
}
result.bitField0_ |= to_bitField0_;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryResponseProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryResponseProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryResponseProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryResponseProto.getDefaultInstance()) return this;
if (other.hasStatus()) {
setStatus(other.getStatus());
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 8: {
status_ = input.readBool();
bitField0_ |= 0x00000001;
break;
} // case 8
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private boolean status_ ;
/**
* <code>optional bool status = 1;</code>
* @return Whether the status field is set.
*/
@java.lang.Override
public boolean hasStatus() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional bool status = 1;</code>
* @return The status.
*/
@java.lang.Override
public boolean getStatus() {
return status_;
}
/**
* <code>optional bool status = 1;</code>
* @param value The status to set.
* @return This builder for chaining.
*/
public Builder setStatus(boolean value) {
status_ = value;
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>optional bool status = 1;</code>
* @return This builder for chaining.
*/
public Builder clearStatus() {
bitField0_ = (bitField0_ & ~0x00000001);
status_ = false;
onChanged();
return this;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.UpdateMountTableEntryResponseProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.UpdateMountTableEntryResponseProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryResponseProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryResponseProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryResponseProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<UpdateMountTableEntryResponseProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<UpdateMountTableEntryResponseProto>() {
@java.lang.Override
public UpdateMountTableEntryResponseProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<UpdateMountTableEntryResponseProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<UpdateMountTableEntryResponseProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.UpdateMountTableEntryResponseProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface RemoveMountTableEntryRequestProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.RemoveMountTableEntryRequestProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* <code>optional string srcPath = 1;</code>
* @return Whether the srcPath field is set.
*/
boolean hasSrcPath();
/**
* <code>optional string srcPath = 1;</code>
* @return The srcPath.
*/
java.lang.String getSrcPath();
/**
* <code>optional string srcPath = 1;</code>
* @return The bytes for srcPath.
*/
org.apache.hadoop.thirdparty.protobuf.ByteString
getSrcPathBytes();
}
/**
* Protobuf type {@code hadoop.hdfs.RemoveMountTableEntryRequestProto}
*/
public static final class RemoveMountTableEntryRequestProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.RemoveMountTableEntryRequestProto)
RemoveMountTableEntryRequestProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use RemoveMountTableEntryRequestProto.newBuilder() to construct.
private RemoveMountTableEntryRequestProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private RemoveMountTableEntryRequestProto() {
srcPath_ = "";
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new RemoveMountTableEntryRequestProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RemoveMountTableEntryRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RemoveMountTableEntryRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryRequestProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryRequestProto.Builder.class);
}
private int bitField0_;
public static final int SRCPATH_FIELD_NUMBER = 1;
@SuppressWarnings("serial")
private volatile java.lang.Object srcPath_ = "";
/**
* <code>optional string srcPath = 1;</code>
* @return Whether the srcPath field is set.
*/
@java.lang.Override
public boolean hasSrcPath() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional string srcPath = 1;</code>
* @return The srcPath.
*/
@java.lang.Override
public java.lang.String getSrcPath() {
java.lang.Object ref = srcPath_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
srcPath_ = s;
}
return s;
}
}
/**
* <code>optional string srcPath = 1;</code>
* @return The bytes for srcPath.
*/
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.ByteString
getSrcPathBytes() {
java.lang.Object ref = srcPath_;
if (ref instanceof java.lang.String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
srcPath_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) != 0)) {
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 1, srcPath_);
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(1, srcPath_);
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryRequestProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryRequestProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryRequestProto) obj;
if (hasSrcPath() != other.hasSrcPath()) return false;
if (hasSrcPath()) {
if (!getSrcPath()
.equals(other.getSrcPath())) return false;
}
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (hasSrcPath()) {
hash = (37 * hash) + SRCPATH_FIELD_NUMBER;
hash = (53 * hash) + getSrcPath().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryRequestProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryRequestProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryRequestProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryRequestProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryRequestProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryRequestProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryRequestProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryRequestProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryRequestProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.RemoveMountTableEntryRequestProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.RemoveMountTableEntryRequestProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryRequestProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RemoveMountTableEntryRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RemoveMountTableEntryRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryRequestProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryRequestProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryRequestProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
srcPath_ = "";
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RemoveMountTableEntryRequestProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryRequestProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryRequestProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryRequestProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryRequestProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryRequestProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryRequestProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryRequestProto(this);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartial0(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryRequestProto result) {
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) != 0)) {
result.srcPath_ = srcPath_;
to_bitField0_ |= 0x00000001;
}
result.bitField0_ |= to_bitField0_;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryRequestProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryRequestProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryRequestProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryRequestProto.getDefaultInstance()) return this;
if (other.hasSrcPath()) {
srcPath_ = other.srcPath_;
bitField0_ |= 0x00000001;
onChanged();
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 10: {
srcPath_ = input.readBytes();
bitField0_ |= 0x00000001;
break;
} // case 10
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private java.lang.Object srcPath_ = "";
/**
* <code>optional string srcPath = 1;</code>
* @return Whether the srcPath field is set.
*/
public boolean hasSrcPath() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional string srcPath = 1;</code>
* @return The srcPath.
*/
public java.lang.String getSrcPath() {
java.lang.Object ref = srcPath_;
if (!(ref instanceof java.lang.String)) {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
srcPath_ = s;
}
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* <code>optional string srcPath = 1;</code>
* @return The bytes for srcPath.
*/
public org.apache.hadoop.thirdparty.protobuf.ByteString
getSrcPathBytes() {
java.lang.Object ref = srcPath_;
if (ref instanceof String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
srcPath_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
/**
* <code>optional string srcPath = 1;</code>
* @param value The srcPath to set.
* @return This builder for chaining.
*/
public Builder setSrcPath(
java.lang.String value) {
if (value == null) { throw new NullPointerException(); }
srcPath_ = value;
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>optional string srcPath = 1;</code>
* @return This builder for chaining.
*/
public Builder clearSrcPath() {
srcPath_ = getDefaultInstance().getSrcPath();
bitField0_ = (bitField0_ & ~0x00000001);
onChanged();
return this;
}
/**
* <code>optional string srcPath = 1;</code>
* @param value The bytes for srcPath to set.
* @return This builder for chaining.
*/
public Builder setSrcPathBytes(
org.apache.hadoop.thirdparty.protobuf.ByteString value) {
if (value == null) { throw new NullPointerException(); }
srcPath_ = value;
bitField0_ |= 0x00000001;
onChanged();
return this;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.RemoveMountTableEntryRequestProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.RemoveMountTableEntryRequestProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryRequestProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryRequestProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryRequestProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<RemoveMountTableEntryRequestProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<RemoveMountTableEntryRequestProto>() {
@java.lang.Override
public RemoveMountTableEntryRequestProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<RemoveMountTableEntryRequestProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<RemoveMountTableEntryRequestProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryRequestProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface RemoveMountTableEntryResponseProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.RemoveMountTableEntryResponseProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* <code>optional bool status = 1;</code>
* @return Whether the status field is set.
*/
boolean hasStatus();
/**
* <code>optional bool status = 1;</code>
* @return The status.
*/
boolean getStatus();
}
/**
* Protobuf type {@code hadoop.hdfs.RemoveMountTableEntryResponseProto}
*/
public static final class RemoveMountTableEntryResponseProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.RemoveMountTableEntryResponseProto)
RemoveMountTableEntryResponseProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use RemoveMountTableEntryResponseProto.newBuilder() to construct.
private RemoveMountTableEntryResponseProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private RemoveMountTableEntryResponseProto() {
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new RemoveMountTableEntryResponseProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RemoveMountTableEntryResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RemoveMountTableEntryResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryResponseProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryResponseProto.Builder.class);
}
private int bitField0_;
public static final int STATUS_FIELD_NUMBER = 1;
private boolean status_ = false;
/**
* <code>optional bool status = 1;</code>
* @return Whether the status field is set.
*/
@java.lang.Override
public boolean hasStatus() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional bool status = 1;</code>
* @return The status.
*/
@java.lang.Override
public boolean getStatus() {
return status_;
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) != 0)) {
output.writeBool(1, status_);
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeBoolSize(1, status_);
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryResponseProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryResponseProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryResponseProto) obj;
if (hasStatus() != other.hasStatus()) return false;
if (hasStatus()) {
if (getStatus()
!= other.getStatus()) return false;
}
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (hasStatus()) {
hash = (37 * hash) + STATUS_FIELD_NUMBER;
hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashBoolean(
getStatus());
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryResponseProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryResponseProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryResponseProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryResponseProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryResponseProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryResponseProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryResponseProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryResponseProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryResponseProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.RemoveMountTableEntryResponseProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.RemoveMountTableEntryResponseProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryResponseProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RemoveMountTableEntryResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RemoveMountTableEntryResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryResponseProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryResponseProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryResponseProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
status_ = false;
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RemoveMountTableEntryResponseProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryResponseProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryResponseProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryResponseProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryResponseProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryResponseProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryResponseProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryResponseProto(this);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartial0(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryResponseProto result) {
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) != 0)) {
result.status_ = status_;
to_bitField0_ |= 0x00000001;
}
result.bitField0_ |= to_bitField0_;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryResponseProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryResponseProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryResponseProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryResponseProto.getDefaultInstance()) return this;
if (other.hasStatus()) {
setStatus(other.getStatus());
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 8: {
status_ = input.readBool();
bitField0_ |= 0x00000001;
break;
} // case 8
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private boolean status_ ;
/**
* <code>optional bool status = 1;</code>
* @return Whether the status field is set.
*/
@java.lang.Override
public boolean hasStatus() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional bool status = 1;</code>
* @return The status.
*/
@java.lang.Override
public boolean getStatus() {
return status_;
}
/**
* <code>optional bool status = 1;</code>
* @param value The status to set.
* @return This builder for chaining.
*/
public Builder setStatus(boolean value) {
status_ = value;
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>optional bool status = 1;</code>
* @return This builder for chaining.
*/
public Builder clearStatus() {
bitField0_ = (bitField0_ & ~0x00000001);
status_ = false;
onChanged();
return this;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.RemoveMountTableEntryResponseProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.RemoveMountTableEntryResponseProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryResponseProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryResponseProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryResponseProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<RemoveMountTableEntryResponseProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<RemoveMountTableEntryResponseProto>() {
@java.lang.Override
public RemoveMountTableEntryResponseProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<RemoveMountTableEntryResponseProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<RemoveMountTableEntryResponseProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RemoveMountTableEntryResponseProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface GetMountTableEntriesRequestProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.GetMountTableEntriesRequestProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* <code>optional string srcPath = 1;</code>
* @return Whether the srcPath field is set.
*/
boolean hasSrcPath();
/**
* <code>optional string srcPath = 1;</code>
* @return The srcPath.
*/
java.lang.String getSrcPath();
/**
* <code>optional string srcPath = 1;</code>
* @return The bytes for srcPath.
*/
org.apache.hadoop.thirdparty.protobuf.ByteString
getSrcPathBytes();
}
/**
* Protobuf type {@code hadoop.hdfs.GetMountTableEntriesRequestProto}
*/
public static final class GetMountTableEntriesRequestProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.GetMountTableEntriesRequestProto)
GetMountTableEntriesRequestProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use GetMountTableEntriesRequestProto.newBuilder() to construct.
private GetMountTableEntriesRequestProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private GetMountTableEntriesRequestProto() {
srcPath_ = "";
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new GetMountTableEntriesRequestProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetMountTableEntriesRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetMountTableEntriesRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesRequestProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesRequestProto.Builder.class);
}
private int bitField0_;
public static final int SRCPATH_FIELD_NUMBER = 1;
@SuppressWarnings("serial")
private volatile java.lang.Object srcPath_ = "";
/**
* <code>optional string srcPath = 1;</code>
* @return Whether the srcPath field is set.
*/
@java.lang.Override
public boolean hasSrcPath() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional string srcPath = 1;</code>
* @return The srcPath.
*/
@java.lang.Override
public java.lang.String getSrcPath() {
java.lang.Object ref = srcPath_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
srcPath_ = s;
}
return s;
}
}
/**
* <code>optional string srcPath = 1;</code>
* @return The bytes for srcPath.
*/
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.ByteString
getSrcPathBytes() {
java.lang.Object ref = srcPath_;
if (ref instanceof java.lang.String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
srcPath_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) != 0)) {
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 1, srcPath_);
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(1, srcPath_);
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesRequestProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesRequestProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesRequestProto) obj;
if (hasSrcPath() != other.hasSrcPath()) return false;
if (hasSrcPath()) {
if (!getSrcPath()
.equals(other.getSrcPath())) return false;
}
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (hasSrcPath()) {
hash = (37 * hash) + SRCPATH_FIELD_NUMBER;
hash = (53 * hash) + getSrcPath().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesRequestProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesRequestProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesRequestProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesRequestProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesRequestProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesRequestProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesRequestProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesRequestProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesRequestProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.GetMountTableEntriesRequestProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.GetMountTableEntriesRequestProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesRequestProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetMountTableEntriesRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetMountTableEntriesRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesRequestProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesRequestProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesRequestProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
srcPath_ = "";
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetMountTableEntriesRequestProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesRequestProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesRequestProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesRequestProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesRequestProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesRequestProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesRequestProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesRequestProto(this);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartial0(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesRequestProto result) {
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) != 0)) {
result.srcPath_ = srcPath_;
to_bitField0_ |= 0x00000001;
}
result.bitField0_ |= to_bitField0_;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesRequestProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesRequestProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesRequestProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesRequestProto.getDefaultInstance()) return this;
if (other.hasSrcPath()) {
srcPath_ = other.srcPath_;
bitField0_ |= 0x00000001;
onChanged();
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 10: {
srcPath_ = input.readBytes();
bitField0_ |= 0x00000001;
break;
} // case 10
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private java.lang.Object srcPath_ = "";
/**
* <code>optional string srcPath = 1;</code>
* @return Whether the srcPath field is set.
*/
public boolean hasSrcPath() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional string srcPath = 1;</code>
* @return The srcPath.
*/
public java.lang.String getSrcPath() {
java.lang.Object ref = srcPath_;
if (!(ref instanceof java.lang.String)) {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
srcPath_ = s;
}
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* <code>optional string srcPath = 1;</code>
* @return The bytes for srcPath.
*/
public org.apache.hadoop.thirdparty.protobuf.ByteString
getSrcPathBytes() {
java.lang.Object ref = srcPath_;
if (ref instanceof String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
srcPath_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
/**
* <code>optional string srcPath = 1;</code>
* @param value The srcPath to set.
* @return This builder for chaining.
*/
public Builder setSrcPath(
java.lang.String value) {
if (value == null) { throw new NullPointerException(); }
srcPath_ = value;
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>optional string srcPath = 1;</code>
* @return This builder for chaining.
*/
public Builder clearSrcPath() {
srcPath_ = getDefaultInstance().getSrcPath();
bitField0_ = (bitField0_ & ~0x00000001);
onChanged();
return this;
}
/**
* <code>optional string srcPath = 1;</code>
* @param value The bytes for srcPath to set.
* @return This builder for chaining.
*/
public Builder setSrcPathBytes(
org.apache.hadoop.thirdparty.protobuf.ByteString value) {
if (value == null) { throw new NullPointerException(); }
srcPath_ = value;
bitField0_ |= 0x00000001;
onChanged();
return this;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.GetMountTableEntriesRequestProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.GetMountTableEntriesRequestProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesRequestProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesRequestProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesRequestProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<GetMountTableEntriesRequestProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<GetMountTableEntriesRequestProto>() {
@java.lang.Override
public GetMountTableEntriesRequestProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<GetMountTableEntriesRequestProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<GetMountTableEntriesRequestProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesRequestProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface GetMountTableEntriesResponseProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.GetMountTableEntriesResponseProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entries = 1;</code>
*/
java.util.List<org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto>
getEntriesList();
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entries = 1;</code>
*/
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto getEntries(int index);
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entries = 1;</code>
*/
int getEntriesCount();
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entries = 1;</code>
*/
java.util.List<? extends org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProtoOrBuilder>
getEntriesOrBuilderList();
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entries = 1;</code>
*/
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProtoOrBuilder getEntriesOrBuilder(
int index);
/**
* <code>optional uint64 timestamp = 2;</code>
* @return Whether the timestamp field is set.
*/
boolean hasTimestamp();
/**
* <code>optional uint64 timestamp = 2;</code>
* @return The timestamp.
*/
long getTimestamp();
}
/**
* Protobuf type {@code hadoop.hdfs.GetMountTableEntriesResponseProto}
*/
public static final class GetMountTableEntriesResponseProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.GetMountTableEntriesResponseProto)
GetMountTableEntriesResponseProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use GetMountTableEntriesResponseProto.newBuilder() to construct.
private GetMountTableEntriesResponseProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private GetMountTableEntriesResponseProto() {
entries_ = java.util.Collections.emptyList();
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new GetMountTableEntriesResponseProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetMountTableEntriesResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetMountTableEntriesResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesResponseProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesResponseProto.Builder.class);
}
private int bitField0_;
public static final int ENTRIES_FIELD_NUMBER = 1;
@SuppressWarnings("serial")
private java.util.List<org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto> entries_;
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entries = 1;</code>
*/
@java.lang.Override
public java.util.List<org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto> getEntriesList() {
return entries_;
}
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entries = 1;</code>
*/
@java.lang.Override
public java.util.List<? extends org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProtoOrBuilder>
getEntriesOrBuilderList() {
return entries_;
}
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entries = 1;</code>
*/
@java.lang.Override
public int getEntriesCount() {
return entries_.size();
}
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entries = 1;</code>
*/
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto getEntries(int index) {
return entries_.get(index);
}
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entries = 1;</code>
*/
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProtoOrBuilder getEntriesOrBuilder(
int index) {
return entries_.get(index);
}
public static final int TIMESTAMP_FIELD_NUMBER = 2;
private long timestamp_ = 0L;
/**
* <code>optional uint64 timestamp = 2;</code>
* @return Whether the timestamp field is set.
*/
@java.lang.Override
public boolean hasTimestamp() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional uint64 timestamp = 2;</code>
* @return The timestamp.
*/
@java.lang.Override
public long getTimestamp() {
return timestamp_;
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
for (int i = 0; i < getEntriesCount(); i++) {
if (!getEntries(i).isInitialized()) {
memoizedIsInitialized = 0;
return false;
}
}
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
for (int i = 0; i < entries_.size(); i++) {
output.writeMessage(1, entries_.get(i));
}
if (((bitField0_ & 0x00000001) != 0)) {
output.writeUInt64(2, timestamp_);
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
for (int i = 0; i < entries_.size(); i++) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeMessageSize(1, entries_.get(i));
}
if (((bitField0_ & 0x00000001) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeUInt64Size(2, timestamp_);
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesResponseProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesResponseProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesResponseProto) obj;
if (!getEntriesList()
.equals(other.getEntriesList())) return false;
if (hasTimestamp() != other.hasTimestamp()) return false;
if (hasTimestamp()) {
if (getTimestamp()
!= other.getTimestamp()) return false;
}
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (getEntriesCount() > 0) {
hash = (37 * hash) + ENTRIES_FIELD_NUMBER;
hash = (53 * hash) + getEntriesList().hashCode();
}
if (hasTimestamp()) {
hash = (37 * hash) + TIMESTAMP_FIELD_NUMBER;
hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashLong(
getTimestamp());
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesResponseProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesResponseProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesResponseProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesResponseProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesResponseProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesResponseProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesResponseProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesResponseProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesResponseProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.GetMountTableEntriesResponseProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.GetMountTableEntriesResponseProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesResponseProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetMountTableEntriesResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetMountTableEntriesResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesResponseProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesResponseProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesResponseProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
if (entriesBuilder_ == null) {
entries_ = java.util.Collections.emptyList();
} else {
entries_ = null;
entriesBuilder_.clear();
}
bitField0_ = (bitField0_ & ~0x00000001);
timestamp_ = 0L;
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetMountTableEntriesResponseProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesResponseProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesResponseProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesResponseProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesResponseProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesResponseProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesResponseProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesResponseProto(this);
buildPartialRepeatedFields(result);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartialRepeatedFields(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesResponseProto result) {
if (entriesBuilder_ == null) {
if (((bitField0_ & 0x00000001) != 0)) {
entries_ = java.util.Collections.unmodifiableList(entries_);
bitField0_ = (bitField0_ & ~0x00000001);
}
result.entries_ = entries_;
} else {
result.entries_ = entriesBuilder_.build();
}
}
private void buildPartial0(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesResponseProto result) {
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000002) != 0)) {
result.timestamp_ = timestamp_;
to_bitField0_ |= 0x00000001;
}
result.bitField0_ |= to_bitField0_;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesResponseProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesResponseProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesResponseProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesResponseProto.getDefaultInstance()) return this;
if (entriesBuilder_ == null) {
if (!other.entries_.isEmpty()) {
if (entries_.isEmpty()) {
entries_ = other.entries_;
bitField0_ = (bitField0_ & ~0x00000001);
} else {
ensureEntriesIsMutable();
entries_.addAll(other.entries_);
}
onChanged();
}
} else {
if (!other.entries_.isEmpty()) {
if (entriesBuilder_.isEmpty()) {
entriesBuilder_.dispose();
entriesBuilder_ = null;
entries_ = other.entries_;
bitField0_ = (bitField0_ & ~0x00000001);
entriesBuilder_ =
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
getEntriesFieldBuilder() : null;
} else {
entriesBuilder_.addAllMessages(other.entries_);
}
}
}
if (other.hasTimestamp()) {
setTimestamp(other.getTimestamp());
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
for (int i = 0; i < getEntriesCount(); i++) {
if (!getEntries(i).isInitialized()) {
return false;
}
}
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 10: {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto m =
input.readMessage(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.PARSER,
extensionRegistry);
if (entriesBuilder_ == null) {
ensureEntriesIsMutable();
entries_.add(m);
} else {
entriesBuilder_.addMessage(m);
}
break;
} // case 10
case 16: {
timestamp_ = input.readUInt64();
bitField0_ |= 0x00000002;
break;
} // case 16
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private java.util.List<org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto> entries_ =
java.util.Collections.emptyList();
private void ensureEntriesIsMutable() {
if (!((bitField0_ & 0x00000001) != 0)) {
entries_ = new java.util.ArrayList<org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto>(entries_);
bitField0_ |= 0x00000001;
}
}
private org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.Builder, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProtoOrBuilder> entriesBuilder_;
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entries = 1;</code>
*/
public java.util.List<org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto> getEntriesList() {
if (entriesBuilder_ == null) {
return java.util.Collections.unmodifiableList(entries_);
} else {
return entriesBuilder_.getMessageList();
}
}
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entries = 1;</code>
*/
public int getEntriesCount() {
if (entriesBuilder_ == null) {
return entries_.size();
} else {
return entriesBuilder_.getCount();
}
}
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entries = 1;</code>
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto getEntries(int index) {
if (entriesBuilder_ == null) {
return entries_.get(index);
} else {
return entriesBuilder_.getMessage(index);
}
}
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entries = 1;</code>
*/
public Builder setEntries(
int index, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto value) {
if (entriesBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
ensureEntriesIsMutable();
entries_.set(index, value);
onChanged();
} else {
entriesBuilder_.setMessage(index, value);
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entries = 1;</code>
*/
public Builder setEntries(
int index, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.Builder builderForValue) {
if (entriesBuilder_ == null) {
ensureEntriesIsMutable();
entries_.set(index, builderForValue.build());
onChanged();
} else {
entriesBuilder_.setMessage(index, builderForValue.build());
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entries = 1;</code>
*/
public Builder addEntries(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto value) {
if (entriesBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
ensureEntriesIsMutable();
entries_.add(value);
onChanged();
} else {
entriesBuilder_.addMessage(value);
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entries = 1;</code>
*/
public Builder addEntries(
int index, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto value) {
if (entriesBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
ensureEntriesIsMutable();
entries_.add(index, value);
onChanged();
} else {
entriesBuilder_.addMessage(index, value);
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entries = 1;</code>
*/
public Builder addEntries(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.Builder builderForValue) {
if (entriesBuilder_ == null) {
ensureEntriesIsMutable();
entries_.add(builderForValue.build());
onChanged();
} else {
entriesBuilder_.addMessage(builderForValue.build());
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entries = 1;</code>
*/
public Builder addEntries(
int index, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.Builder builderForValue) {
if (entriesBuilder_ == null) {
ensureEntriesIsMutable();
entries_.add(index, builderForValue.build());
onChanged();
} else {
entriesBuilder_.addMessage(index, builderForValue.build());
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entries = 1;</code>
*/
public Builder addAllEntries(
java.lang.Iterable<? extends org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto> values) {
if (entriesBuilder_ == null) {
ensureEntriesIsMutable();
org.apache.hadoop.thirdparty.protobuf.AbstractMessageLite.Builder.addAll(
values, entries_);
onChanged();
} else {
entriesBuilder_.addAllMessages(values);
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entries = 1;</code>
*/
public Builder clearEntries() {
if (entriesBuilder_ == null) {
entries_ = java.util.Collections.emptyList();
bitField0_ = (bitField0_ & ~0x00000001);
onChanged();
} else {
entriesBuilder_.clear();
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entries = 1;</code>
*/
public Builder removeEntries(int index) {
if (entriesBuilder_ == null) {
ensureEntriesIsMutable();
entries_.remove(index);
onChanged();
} else {
entriesBuilder_.remove(index);
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entries = 1;</code>
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.Builder getEntriesBuilder(
int index) {
return getEntriesFieldBuilder().getBuilder(index);
}
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entries = 1;</code>
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProtoOrBuilder getEntriesOrBuilder(
int index) {
if (entriesBuilder_ == null) {
return entries_.get(index); } else {
return entriesBuilder_.getMessageOrBuilder(index);
}
}
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entries = 1;</code>
*/
public java.util.List<? extends org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProtoOrBuilder>
getEntriesOrBuilderList() {
if (entriesBuilder_ != null) {
return entriesBuilder_.getMessageOrBuilderList();
} else {
return java.util.Collections.unmodifiableList(entries_);
}
}
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entries = 1;</code>
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.Builder addEntriesBuilder() {
return getEntriesFieldBuilder().addBuilder(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.getDefaultInstance());
}
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entries = 1;</code>
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.Builder addEntriesBuilder(
int index) {
return getEntriesFieldBuilder().addBuilder(
index, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.getDefaultInstance());
}
/**
* <code>repeated .hadoop.hdfs.MountTableRecordProto entries = 1;</code>
*/
public java.util.List<org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.Builder>
getEntriesBuilderList() {
return getEntriesFieldBuilder().getBuilderList();
}
private org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.Builder, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProtoOrBuilder>
getEntriesFieldBuilder() {
if (entriesBuilder_ == null) {
entriesBuilder_ = new org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProto.Builder, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.MountTableRecordProtoOrBuilder>(
entries_,
((bitField0_ & 0x00000001) != 0),
getParentForChildren(),
isClean());
entries_ = null;
}
return entriesBuilder_;
}
private long timestamp_ ;
/**
* <code>optional uint64 timestamp = 2;</code>
* @return Whether the timestamp field is set.
*/
@java.lang.Override
public boolean hasTimestamp() {
return ((bitField0_ & 0x00000002) != 0);
}
/**
* <code>optional uint64 timestamp = 2;</code>
* @return The timestamp.
*/
@java.lang.Override
public long getTimestamp() {
return timestamp_;
}
/**
* <code>optional uint64 timestamp = 2;</code>
* @param value The timestamp to set.
* @return This builder for chaining.
*/
public Builder setTimestamp(long value) {
timestamp_ = value;
bitField0_ |= 0x00000002;
onChanged();
return this;
}
/**
* <code>optional uint64 timestamp = 2;</code>
* @return This builder for chaining.
*/
public Builder clearTimestamp() {
bitField0_ = (bitField0_ & ~0x00000002);
timestamp_ = 0L;
onChanged();
return this;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.GetMountTableEntriesResponseProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.GetMountTableEntriesResponseProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesResponseProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesResponseProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesResponseProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<GetMountTableEntriesResponseProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<GetMountTableEntriesResponseProto>() {
@java.lang.Override
public GetMountTableEntriesResponseProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<GetMountTableEntriesResponseProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<GetMountTableEntriesResponseProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetMountTableEntriesResponseProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface GetDestinationRequestProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.GetDestinationRequestProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* <code>optional string srcPath = 1;</code>
* @return Whether the srcPath field is set.
*/
boolean hasSrcPath();
/**
* <code>optional string srcPath = 1;</code>
* @return The srcPath.
*/
java.lang.String getSrcPath();
/**
* <code>optional string srcPath = 1;</code>
* @return The bytes for srcPath.
*/
org.apache.hadoop.thirdparty.protobuf.ByteString
getSrcPathBytes();
}
/**
* Protobuf type {@code hadoop.hdfs.GetDestinationRequestProto}
*/
public static final class GetDestinationRequestProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.GetDestinationRequestProto)
GetDestinationRequestProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use GetDestinationRequestProto.newBuilder() to construct.
private GetDestinationRequestProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private GetDestinationRequestProto() {
srcPath_ = "";
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new GetDestinationRequestProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetDestinationRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetDestinationRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationRequestProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationRequestProto.Builder.class);
}
private int bitField0_;
public static final int SRCPATH_FIELD_NUMBER = 1;
@SuppressWarnings("serial")
private volatile java.lang.Object srcPath_ = "";
/**
* <code>optional string srcPath = 1;</code>
* @return Whether the srcPath field is set.
*/
@java.lang.Override
public boolean hasSrcPath() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional string srcPath = 1;</code>
* @return The srcPath.
*/
@java.lang.Override
public java.lang.String getSrcPath() {
java.lang.Object ref = srcPath_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
srcPath_ = s;
}
return s;
}
}
/**
* <code>optional string srcPath = 1;</code>
* @return The bytes for srcPath.
*/
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.ByteString
getSrcPathBytes() {
java.lang.Object ref = srcPath_;
if (ref instanceof java.lang.String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
srcPath_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) != 0)) {
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 1, srcPath_);
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(1, srcPath_);
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationRequestProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationRequestProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationRequestProto) obj;
if (hasSrcPath() != other.hasSrcPath()) return false;
if (hasSrcPath()) {
if (!getSrcPath()
.equals(other.getSrcPath())) return false;
}
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (hasSrcPath()) {
hash = (37 * hash) + SRCPATH_FIELD_NUMBER;
hash = (53 * hash) + getSrcPath().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationRequestProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationRequestProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationRequestProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationRequestProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationRequestProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationRequestProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationRequestProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationRequestProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationRequestProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.GetDestinationRequestProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.GetDestinationRequestProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationRequestProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetDestinationRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetDestinationRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationRequestProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationRequestProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationRequestProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
srcPath_ = "";
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetDestinationRequestProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationRequestProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationRequestProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationRequestProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationRequestProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationRequestProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationRequestProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationRequestProto(this);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartial0(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationRequestProto result) {
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) != 0)) {
result.srcPath_ = srcPath_;
to_bitField0_ |= 0x00000001;
}
result.bitField0_ |= to_bitField0_;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationRequestProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationRequestProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationRequestProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationRequestProto.getDefaultInstance()) return this;
if (other.hasSrcPath()) {
srcPath_ = other.srcPath_;
bitField0_ |= 0x00000001;
onChanged();
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 10: {
srcPath_ = input.readBytes();
bitField0_ |= 0x00000001;
break;
} // case 10
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private java.lang.Object srcPath_ = "";
/**
* <code>optional string srcPath = 1;</code>
* @return Whether the srcPath field is set.
*/
public boolean hasSrcPath() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional string srcPath = 1;</code>
* @return The srcPath.
*/
public java.lang.String getSrcPath() {
java.lang.Object ref = srcPath_;
if (!(ref instanceof java.lang.String)) {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
srcPath_ = s;
}
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* <code>optional string srcPath = 1;</code>
* @return The bytes for srcPath.
*/
public org.apache.hadoop.thirdparty.protobuf.ByteString
getSrcPathBytes() {
java.lang.Object ref = srcPath_;
if (ref instanceof String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
srcPath_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
/**
* <code>optional string srcPath = 1;</code>
* @param value The srcPath to set.
* @return This builder for chaining.
*/
public Builder setSrcPath(
java.lang.String value) {
if (value == null) { throw new NullPointerException(); }
srcPath_ = value;
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>optional string srcPath = 1;</code>
* @return This builder for chaining.
*/
public Builder clearSrcPath() {
srcPath_ = getDefaultInstance().getSrcPath();
bitField0_ = (bitField0_ & ~0x00000001);
onChanged();
return this;
}
/**
* <code>optional string srcPath = 1;</code>
* @param value The bytes for srcPath to set.
* @return This builder for chaining.
*/
public Builder setSrcPathBytes(
org.apache.hadoop.thirdparty.protobuf.ByteString value) {
if (value == null) { throw new NullPointerException(); }
srcPath_ = value;
bitField0_ |= 0x00000001;
onChanged();
return this;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.GetDestinationRequestProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.GetDestinationRequestProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationRequestProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationRequestProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationRequestProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<GetDestinationRequestProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<GetDestinationRequestProto>() {
@java.lang.Override
public GetDestinationRequestProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<GetDestinationRequestProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<GetDestinationRequestProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationRequestProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface GetDestinationResponseProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.GetDestinationResponseProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* <code>repeated string destinations = 1;</code>
* @return A list containing the destinations.
*/
java.util.List<java.lang.String>
getDestinationsList();
/**
* <code>repeated string destinations = 1;</code>
* @return The count of destinations.
*/
int getDestinationsCount();
/**
* <code>repeated string destinations = 1;</code>
* @param index The index of the element to return.
* @return The destinations at the given index.
*/
java.lang.String getDestinations(int index);
/**
* <code>repeated string destinations = 1;</code>
* @param index The index of the value to return.
* @return The bytes of the destinations at the given index.
*/
org.apache.hadoop.thirdparty.protobuf.ByteString
getDestinationsBytes(int index);
}
/**
* Protobuf type {@code hadoop.hdfs.GetDestinationResponseProto}
*/
public static final class GetDestinationResponseProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.GetDestinationResponseProto)
GetDestinationResponseProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use GetDestinationResponseProto.newBuilder() to construct.
private GetDestinationResponseProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private GetDestinationResponseProto() {
destinations_ =
org.apache.hadoop.thirdparty.protobuf.LazyStringArrayList.emptyList();
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new GetDestinationResponseProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetDestinationResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetDestinationResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationResponseProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationResponseProto.Builder.class);
}
public static final int DESTINATIONS_FIELD_NUMBER = 1;
@SuppressWarnings("serial")
private org.apache.hadoop.thirdparty.protobuf.LazyStringArrayList destinations_ =
org.apache.hadoop.thirdparty.protobuf.LazyStringArrayList.emptyList();
/**
* <code>repeated string destinations = 1;</code>
* @return A list containing the destinations.
*/
public org.apache.hadoop.thirdparty.protobuf.ProtocolStringList
getDestinationsList() {
return destinations_;
}
/**
* <code>repeated string destinations = 1;</code>
* @return The count of destinations.
*/
public int getDestinationsCount() {
return destinations_.size();
}
/**
* <code>repeated string destinations = 1;</code>
* @param index The index of the element to return.
* @return The destinations at the given index.
*/
public java.lang.String getDestinations(int index) {
return destinations_.get(index);
}
/**
* <code>repeated string destinations = 1;</code>
* @param index The index of the value to return.
* @return The bytes of the destinations at the given index.
*/
public org.apache.hadoop.thirdparty.protobuf.ByteString
getDestinationsBytes(int index) {
return destinations_.getByteString(index);
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
for (int i = 0; i < destinations_.size(); i++) {
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 1, destinations_.getRaw(i));
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
{
int dataSize = 0;
for (int i = 0; i < destinations_.size(); i++) {
dataSize += computeStringSizeNoTag(destinations_.getRaw(i));
}
size += dataSize;
size += 1 * getDestinationsList().size();
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationResponseProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationResponseProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationResponseProto) obj;
if (!getDestinationsList()
.equals(other.getDestinationsList())) return false;
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (getDestinationsCount() > 0) {
hash = (37 * hash) + DESTINATIONS_FIELD_NUMBER;
hash = (53 * hash) + getDestinationsList().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationResponseProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationResponseProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationResponseProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationResponseProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationResponseProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationResponseProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationResponseProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationResponseProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationResponseProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.GetDestinationResponseProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.GetDestinationResponseProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationResponseProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetDestinationResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetDestinationResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationResponseProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationResponseProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationResponseProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
destinations_ =
org.apache.hadoop.thirdparty.protobuf.LazyStringArrayList.emptyList();
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetDestinationResponseProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationResponseProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationResponseProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationResponseProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationResponseProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationResponseProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationResponseProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationResponseProto(this);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartial0(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationResponseProto result) {
int from_bitField0_ = bitField0_;
if (((from_bitField0_ & 0x00000001) != 0)) {
destinations_.makeImmutable();
result.destinations_ = destinations_;
}
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationResponseProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationResponseProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationResponseProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationResponseProto.getDefaultInstance()) return this;
if (!other.destinations_.isEmpty()) {
if (destinations_.isEmpty()) {
destinations_ = other.destinations_;
bitField0_ |= 0x00000001;
} else {
ensureDestinationsIsMutable();
destinations_.addAll(other.destinations_);
}
onChanged();
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 10: {
org.apache.hadoop.thirdparty.protobuf.ByteString bs = input.readBytes();
ensureDestinationsIsMutable();
destinations_.add(bs);
break;
} // case 10
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private org.apache.hadoop.thirdparty.protobuf.LazyStringArrayList destinations_ =
org.apache.hadoop.thirdparty.protobuf.LazyStringArrayList.emptyList();
private void ensureDestinationsIsMutable() {
if (!destinations_.isModifiable()) {
destinations_ = new org.apache.hadoop.thirdparty.protobuf.LazyStringArrayList(destinations_);
}
bitField0_ |= 0x00000001;
}
/**
* <code>repeated string destinations = 1;</code>
* @return A list containing the destinations.
*/
public org.apache.hadoop.thirdparty.protobuf.ProtocolStringList
getDestinationsList() {
destinations_.makeImmutable();
return destinations_;
}
/**
* <code>repeated string destinations = 1;</code>
* @return The count of destinations.
*/
public int getDestinationsCount() {
return destinations_.size();
}
/**
* <code>repeated string destinations = 1;</code>
* @param index The index of the element to return.
* @return The destinations at the given index.
*/
public java.lang.String getDestinations(int index) {
return destinations_.get(index);
}
/**
* <code>repeated string destinations = 1;</code>
* @param index The index of the value to return.
* @return The bytes of the destinations at the given index.
*/
public org.apache.hadoop.thirdparty.protobuf.ByteString
getDestinationsBytes(int index) {
return destinations_.getByteString(index);
}
/**
* <code>repeated string destinations = 1;</code>
* @param index The index to set the value at.
* @param value The destinations to set.
* @return This builder for chaining.
*/
public Builder setDestinations(
int index, java.lang.String value) {
if (value == null) { throw new NullPointerException(); }
ensureDestinationsIsMutable();
destinations_.set(index, value);
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>repeated string destinations = 1;</code>
* @param value The destinations to add.
* @return This builder for chaining.
*/
public Builder addDestinations(
java.lang.String value) {
if (value == null) { throw new NullPointerException(); }
ensureDestinationsIsMutable();
destinations_.add(value);
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>repeated string destinations = 1;</code>
* @param values The destinations to add.
* @return This builder for chaining.
*/
public Builder addAllDestinations(
java.lang.Iterable<java.lang.String> values) {
ensureDestinationsIsMutable();
org.apache.hadoop.thirdparty.protobuf.AbstractMessageLite.Builder.addAll(
values, destinations_);
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>repeated string destinations = 1;</code>
* @return This builder for chaining.
*/
public Builder clearDestinations() {
destinations_ =
org.apache.hadoop.thirdparty.protobuf.LazyStringArrayList.emptyList();
bitField0_ = (bitField0_ & ~0x00000001);;
onChanged();
return this;
}
/**
* <code>repeated string destinations = 1;</code>
* @param value The bytes of the destinations to add.
* @return This builder for chaining.
*/
public Builder addDestinationsBytes(
org.apache.hadoop.thirdparty.protobuf.ByteString value) {
if (value == null) { throw new NullPointerException(); }
ensureDestinationsIsMutable();
destinations_.add(value);
bitField0_ |= 0x00000001;
onChanged();
return this;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.GetDestinationResponseProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.GetDestinationResponseProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationResponseProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationResponseProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationResponseProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<GetDestinationResponseProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<GetDestinationResponseProto>() {
@java.lang.Override
public GetDestinationResponseProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<GetDestinationResponseProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<GetDestinationResponseProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDestinationResponseProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface StateStoreVersionRecordProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.StateStoreVersionRecordProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* <code>optional uint64 membershipVersion = 1;</code>
* @return Whether the membershipVersion field is set.
*/
boolean hasMembershipVersion();
/**
* <code>optional uint64 membershipVersion = 1;</code>
* @return The membershipVersion.
*/
long getMembershipVersion();
/**
* <code>optional uint64 mountTableVersion = 2;</code>
* @return Whether the mountTableVersion field is set.
*/
boolean hasMountTableVersion();
/**
* <code>optional uint64 mountTableVersion = 2;</code>
* @return The mountTableVersion.
*/
long getMountTableVersion();
}
/**
* Protobuf type {@code hadoop.hdfs.StateStoreVersionRecordProto}
*/
public static final class StateStoreVersionRecordProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.StateStoreVersionRecordProto)
StateStoreVersionRecordProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use StateStoreVersionRecordProto.newBuilder() to construct.
private StateStoreVersionRecordProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private StateStoreVersionRecordProto() {
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new StateStoreVersionRecordProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_StateStoreVersionRecordProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_StateStoreVersionRecordProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto.Builder.class);
}
private int bitField0_;
public static final int MEMBERSHIPVERSION_FIELD_NUMBER = 1;
private long membershipVersion_ = 0L;
/**
* <code>optional uint64 membershipVersion = 1;</code>
* @return Whether the membershipVersion field is set.
*/
@java.lang.Override
public boolean hasMembershipVersion() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional uint64 membershipVersion = 1;</code>
* @return The membershipVersion.
*/
@java.lang.Override
public long getMembershipVersion() {
return membershipVersion_;
}
public static final int MOUNTTABLEVERSION_FIELD_NUMBER = 2;
private long mountTableVersion_ = 0L;
/**
* <code>optional uint64 mountTableVersion = 2;</code>
* @return Whether the mountTableVersion field is set.
*/
@java.lang.Override
public boolean hasMountTableVersion() {
return ((bitField0_ & 0x00000002) != 0);
}
/**
* <code>optional uint64 mountTableVersion = 2;</code>
* @return The mountTableVersion.
*/
@java.lang.Override
public long getMountTableVersion() {
return mountTableVersion_;
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) != 0)) {
output.writeUInt64(1, membershipVersion_);
}
if (((bitField0_ & 0x00000002) != 0)) {
output.writeUInt64(2, mountTableVersion_);
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeUInt64Size(1, membershipVersion_);
}
if (((bitField0_ & 0x00000002) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeUInt64Size(2, mountTableVersion_);
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto) obj;
if (hasMembershipVersion() != other.hasMembershipVersion()) return false;
if (hasMembershipVersion()) {
if (getMembershipVersion()
!= other.getMembershipVersion()) return false;
}
if (hasMountTableVersion() != other.hasMountTableVersion()) return false;
if (hasMountTableVersion()) {
if (getMountTableVersion()
!= other.getMountTableVersion()) return false;
}
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (hasMembershipVersion()) {
hash = (37 * hash) + MEMBERSHIPVERSION_FIELD_NUMBER;
hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashLong(
getMembershipVersion());
}
if (hasMountTableVersion()) {
hash = (37 * hash) + MOUNTTABLEVERSION_FIELD_NUMBER;
hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashLong(
getMountTableVersion());
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.StateStoreVersionRecordProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.StateStoreVersionRecordProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_StateStoreVersionRecordProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_StateStoreVersionRecordProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
membershipVersion_ = 0L;
mountTableVersion_ = 0L;
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_StateStoreVersionRecordProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto(this);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartial0(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto result) {
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) != 0)) {
result.membershipVersion_ = membershipVersion_;
to_bitField0_ |= 0x00000001;
}
if (((from_bitField0_ & 0x00000002) != 0)) {
result.mountTableVersion_ = mountTableVersion_;
to_bitField0_ |= 0x00000002;
}
result.bitField0_ |= to_bitField0_;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto.getDefaultInstance()) return this;
if (other.hasMembershipVersion()) {
setMembershipVersion(other.getMembershipVersion());
}
if (other.hasMountTableVersion()) {
setMountTableVersion(other.getMountTableVersion());
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 8: {
membershipVersion_ = input.readUInt64();
bitField0_ |= 0x00000001;
break;
} // case 8
case 16: {
mountTableVersion_ = input.readUInt64();
bitField0_ |= 0x00000002;
break;
} // case 16
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private long membershipVersion_ ;
/**
* <code>optional uint64 membershipVersion = 1;</code>
* @return Whether the membershipVersion field is set.
*/
@java.lang.Override
public boolean hasMembershipVersion() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional uint64 membershipVersion = 1;</code>
* @return The membershipVersion.
*/
@java.lang.Override
public long getMembershipVersion() {
return membershipVersion_;
}
/**
* <code>optional uint64 membershipVersion = 1;</code>
* @param value The membershipVersion to set.
* @return This builder for chaining.
*/
public Builder setMembershipVersion(long value) {
membershipVersion_ = value;
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>optional uint64 membershipVersion = 1;</code>
* @return This builder for chaining.
*/
public Builder clearMembershipVersion() {
bitField0_ = (bitField0_ & ~0x00000001);
membershipVersion_ = 0L;
onChanged();
return this;
}
private long mountTableVersion_ ;
/**
* <code>optional uint64 mountTableVersion = 2;</code>
* @return Whether the mountTableVersion field is set.
*/
@java.lang.Override
public boolean hasMountTableVersion() {
return ((bitField0_ & 0x00000002) != 0);
}
/**
* <code>optional uint64 mountTableVersion = 2;</code>
* @return The mountTableVersion.
*/
@java.lang.Override
public long getMountTableVersion() {
return mountTableVersion_;
}
/**
* <code>optional uint64 mountTableVersion = 2;</code>
* @param value The mountTableVersion to set.
* @return This builder for chaining.
*/
public Builder setMountTableVersion(long value) {
mountTableVersion_ = value;
bitField0_ |= 0x00000002;
onChanged();
return this;
}
/**
* <code>optional uint64 mountTableVersion = 2;</code>
* @return This builder for chaining.
*/
public Builder clearMountTableVersion() {
bitField0_ = (bitField0_ & ~0x00000002);
mountTableVersion_ = 0L;
onChanged();
return this;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.StateStoreVersionRecordProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.StateStoreVersionRecordProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<StateStoreVersionRecordProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<StateStoreVersionRecordProto>() {
@java.lang.Override
public StateStoreVersionRecordProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<StateStoreVersionRecordProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<StateStoreVersionRecordProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface RouterRecordProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.RouterRecordProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* <code>optional uint64 dateCreated = 1;</code>
* @return Whether the dateCreated field is set.
*/
boolean hasDateCreated();
/**
* <code>optional uint64 dateCreated = 1;</code>
* @return The dateCreated.
*/
long getDateCreated();
/**
* <code>optional uint64 dateModified = 2;</code>
* @return Whether the dateModified field is set.
*/
boolean hasDateModified();
/**
* <code>optional uint64 dateModified = 2;</code>
* @return The dateModified.
*/
long getDateModified();
/**
* <code>optional string address = 3;</code>
* @return Whether the address field is set.
*/
boolean hasAddress();
/**
* <code>optional string address = 3;</code>
* @return The address.
*/
java.lang.String getAddress();
/**
* <code>optional string address = 3;</code>
* @return The bytes for address.
*/
org.apache.hadoop.thirdparty.protobuf.ByteString
getAddressBytes();
/**
* <code>optional string status = 4;</code>
* @return Whether the status field is set.
*/
boolean hasStatus();
/**
* <code>optional string status = 4;</code>
* @return The status.
*/
java.lang.String getStatus();
/**
* <code>optional string status = 4;</code>
* @return The bytes for status.
*/
org.apache.hadoop.thirdparty.protobuf.ByteString
getStatusBytes();
/**
* <code>optional .hadoop.hdfs.StateStoreVersionRecordProto stateStoreVersion = 5;</code>
* @return Whether the stateStoreVersion field is set.
*/
boolean hasStateStoreVersion();
/**
* <code>optional .hadoop.hdfs.StateStoreVersionRecordProto stateStoreVersion = 5;</code>
* @return The stateStoreVersion.
*/
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto getStateStoreVersion();
/**
* <code>optional .hadoop.hdfs.StateStoreVersionRecordProto stateStoreVersion = 5;</code>
*/
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProtoOrBuilder getStateStoreVersionOrBuilder();
/**
* <code>optional string version = 6;</code>
* @return Whether the version field is set.
*/
boolean hasVersion();
/**
* <code>optional string version = 6;</code>
* @return The version.
*/
java.lang.String getVersion();
/**
* <code>optional string version = 6;</code>
* @return The bytes for version.
*/
org.apache.hadoop.thirdparty.protobuf.ByteString
getVersionBytes();
/**
* <code>optional string compileInfo = 7;</code>
* @return Whether the compileInfo field is set.
*/
boolean hasCompileInfo();
/**
* <code>optional string compileInfo = 7;</code>
* @return The compileInfo.
*/
java.lang.String getCompileInfo();
/**
* <code>optional string compileInfo = 7;</code>
* @return The bytes for compileInfo.
*/
org.apache.hadoop.thirdparty.protobuf.ByteString
getCompileInfoBytes();
/**
* <code>optional uint64 dateStarted = 8;</code>
* @return Whether the dateStarted field is set.
*/
boolean hasDateStarted();
/**
* <code>optional uint64 dateStarted = 8;</code>
* @return The dateStarted.
*/
long getDateStarted();
/**
* <code>optional string adminAddress = 9;</code>
* @return Whether the adminAddress field is set.
*/
boolean hasAdminAddress();
/**
* <code>optional string adminAddress = 9;</code>
* @return The adminAddress.
*/
java.lang.String getAdminAddress();
/**
* <code>optional string adminAddress = 9;</code>
* @return The bytes for adminAddress.
*/
org.apache.hadoop.thirdparty.protobuf.ByteString
getAdminAddressBytes();
}
/**
* Protobuf type {@code hadoop.hdfs.RouterRecordProto}
*/
public static final class RouterRecordProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.RouterRecordProto)
RouterRecordProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use RouterRecordProto.newBuilder() to construct.
private RouterRecordProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private RouterRecordProto() {
address_ = "";
status_ = "";
version_ = "";
compileInfo_ = "";
adminAddress_ = "";
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new RouterRecordProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RouterRecordProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RouterRecordProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto.Builder.class);
}
private int bitField0_;
public static final int DATECREATED_FIELD_NUMBER = 1;
private long dateCreated_ = 0L;
/**
* <code>optional uint64 dateCreated = 1;</code>
* @return Whether the dateCreated field is set.
*/
@java.lang.Override
public boolean hasDateCreated() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional uint64 dateCreated = 1;</code>
* @return The dateCreated.
*/
@java.lang.Override
public long getDateCreated() {
return dateCreated_;
}
public static final int DATEMODIFIED_FIELD_NUMBER = 2;
private long dateModified_ = 0L;
/**
* <code>optional uint64 dateModified = 2;</code>
* @return Whether the dateModified field is set.
*/
@java.lang.Override
public boolean hasDateModified() {
return ((bitField0_ & 0x00000002) != 0);
}
/**
* <code>optional uint64 dateModified = 2;</code>
* @return The dateModified.
*/
@java.lang.Override
public long getDateModified() {
return dateModified_;
}
public static final int ADDRESS_FIELD_NUMBER = 3;
@SuppressWarnings("serial")
private volatile java.lang.Object address_ = "";
/**
* <code>optional string address = 3;</code>
* @return Whether the address field is set.
*/
@java.lang.Override
public boolean hasAddress() {
return ((bitField0_ & 0x00000004) != 0);
}
/**
* <code>optional string address = 3;</code>
* @return The address.
*/
@java.lang.Override
public java.lang.String getAddress() {
java.lang.Object ref = address_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
address_ = s;
}
return s;
}
}
/**
* <code>optional string address = 3;</code>
* @return The bytes for address.
*/
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.ByteString
getAddressBytes() {
java.lang.Object ref = address_;
if (ref instanceof java.lang.String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
address_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
public static final int STATUS_FIELD_NUMBER = 4;
@SuppressWarnings("serial")
private volatile java.lang.Object status_ = "";
/**
* <code>optional string status = 4;</code>
* @return Whether the status field is set.
*/
@java.lang.Override
public boolean hasStatus() {
return ((bitField0_ & 0x00000008) != 0);
}
/**
* <code>optional string status = 4;</code>
* @return The status.
*/
@java.lang.Override
public java.lang.String getStatus() {
java.lang.Object ref = status_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
status_ = s;
}
return s;
}
}
/**
* <code>optional string status = 4;</code>
* @return The bytes for status.
*/
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.ByteString
getStatusBytes() {
java.lang.Object ref = status_;
if (ref instanceof java.lang.String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
status_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
public static final int STATESTOREVERSION_FIELD_NUMBER = 5;
private org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto stateStoreVersion_;
/**
* <code>optional .hadoop.hdfs.StateStoreVersionRecordProto stateStoreVersion = 5;</code>
* @return Whether the stateStoreVersion field is set.
*/
@java.lang.Override
public boolean hasStateStoreVersion() {
return ((bitField0_ & 0x00000010) != 0);
}
/**
* <code>optional .hadoop.hdfs.StateStoreVersionRecordProto stateStoreVersion = 5;</code>
* @return The stateStoreVersion.
*/
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto getStateStoreVersion() {
return stateStoreVersion_ == null ? org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto.getDefaultInstance() : stateStoreVersion_;
}
/**
* <code>optional .hadoop.hdfs.StateStoreVersionRecordProto stateStoreVersion = 5;</code>
*/
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProtoOrBuilder getStateStoreVersionOrBuilder() {
return stateStoreVersion_ == null ? org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto.getDefaultInstance() : stateStoreVersion_;
}
public static final int VERSION_FIELD_NUMBER = 6;
@SuppressWarnings("serial")
private volatile java.lang.Object version_ = "";
/**
* <code>optional string version = 6;</code>
* @return Whether the version field is set.
*/
@java.lang.Override
public boolean hasVersion() {
return ((bitField0_ & 0x00000020) != 0);
}
/**
* <code>optional string version = 6;</code>
* @return The version.
*/
@java.lang.Override
public java.lang.String getVersion() {
java.lang.Object ref = version_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
version_ = s;
}
return s;
}
}
/**
* <code>optional string version = 6;</code>
* @return The bytes for version.
*/
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.ByteString
getVersionBytes() {
java.lang.Object ref = version_;
if (ref instanceof java.lang.String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
version_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
public static final int COMPILEINFO_FIELD_NUMBER = 7;
@SuppressWarnings("serial")
private volatile java.lang.Object compileInfo_ = "";
/**
* <code>optional string compileInfo = 7;</code>
* @return Whether the compileInfo field is set.
*/
@java.lang.Override
public boolean hasCompileInfo() {
return ((bitField0_ & 0x00000040) != 0);
}
/**
* <code>optional string compileInfo = 7;</code>
* @return The compileInfo.
*/
@java.lang.Override
public java.lang.String getCompileInfo() {
java.lang.Object ref = compileInfo_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
compileInfo_ = s;
}
return s;
}
}
/**
* <code>optional string compileInfo = 7;</code>
* @return The bytes for compileInfo.
*/
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.ByteString
getCompileInfoBytes() {
java.lang.Object ref = compileInfo_;
if (ref instanceof java.lang.String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
compileInfo_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
public static final int DATESTARTED_FIELD_NUMBER = 8;
private long dateStarted_ = 0L;
/**
* <code>optional uint64 dateStarted = 8;</code>
* @return Whether the dateStarted field is set.
*/
@java.lang.Override
public boolean hasDateStarted() {
return ((bitField0_ & 0x00000080) != 0);
}
/**
* <code>optional uint64 dateStarted = 8;</code>
* @return The dateStarted.
*/
@java.lang.Override
public long getDateStarted() {
return dateStarted_;
}
public static final int ADMINADDRESS_FIELD_NUMBER = 9;
@SuppressWarnings("serial")
private volatile java.lang.Object adminAddress_ = "";
/**
* <code>optional string adminAddress = 9;</code>
* @return Whether the adminAddress field is set.
*/
@java.lang.Override
public boolean hasAdminAddress() {
return ((bitField0_ & 0x00000100) != 0);
}
/**
* <code>optional string adminAddress = 9;</code>
* @return The adminAddress.
*/
@java.lang.Override
public java.lang.String getAdminAddress() {
java.lang.Object ref = adminAddress_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
adminAddress_ = s;
}
return s;
}
}
/**
* <code>optional string adminAddress = 9;</code>
* @return The bytes for adminAddress.
*/
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.ByteString
getAdminAddressBytes() {
java.lang.Object ref = adminAddress_;
if (ref instanceof java.lang.String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
adminAddress_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) != 0)) {
output.writeUInt64(1, dateCreated_);
}
if (((bitField0_ & 0x00000002) != 0)) {
output.writeUInt64(2, dateModified_);
}
if (((bitField0_ & 0x00000004) != 0)) {
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 3, address_);
}
if (((bitField0_ & 0x00000008) != 0)) {
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 4, status_);
}
if (((bitField0_ & 0x00000010) != 0)) {
output.writeMessage(5, getStateStoreVersion());
}
if (((bitField0_ & 0x00000020) != 0)) {
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 6, version_);
}
if (((bitField0_ & 0x00000040) != 0)) {
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 7, compileInfo_);
}
if (((bitField0_ & 0x00000080) != 0)) {
output.writeUInt64(8, dateStarted_);
}
if (((bitField0_ & 0x00000100) != 0)) {
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 9, adminAddress_);
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeUInt64Size(1, dateCreated_);
}
if (((bitField0_ & 0x00000002) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeUInt64Size(2, dateModified_);
}
if (((bitField0_ & 0x00000004) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(3, address_);
}
if (((bitField0_ & 0x00000008) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(4, status_);
}
if (((bitField0_ & 0x00000010) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeMessageSize(5, getStateStoreVersion());
}
if (((bitField0_ & 0x00000020) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(6, version_);
}
if (((bitField0_ & 0x00000040) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(7, compileInfo_);
}
if (((bitField0_ & 0x00000080) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeUInt64Size(8, dateStarted_);
}
if (((bitField0_ & 0x00000100) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(9, adminAddress_);
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto) obj;
if (hasDateCreated() != other.hasDateCreated()) return false;
if (hasDateCreated()) {
if (getDateCreated()
!= other.getDateCreated()) return false;
}
if (hasDateModified() != other.hasDateModified()) return false;
if (hasDateModified()) {
if (getDateModified()
!= other.getDateModified()) return false;
}
if (hasAddress() != other.hasAddress()) return false;
if (hasAddress()) {
if (!getAddress()
.equals(other.getAddress())) return false;
}
if (hasStatus() != other.hasStatus()) return false;
if (hasStatus()) {
if (!getStatus()
.equals(other.getStatus())) return false;
}
if (hasStateStoreVersion() != other.hasStateStoreVersion()) return false;
if (hasStateStoreVersion()) {
if (!getStateStoreVersion()
.equals(other.getStateStoreVersion())) return false;
}
if (hasVersion() != other.hasVersion()) return false;
if (hasVersion()) {
if (!getVersion()
.equals(other.getVersion())) return false;
}
if (hasCompileInfo() != other.hasCompileInfo()) return false;
if (hasCompileInfo()) {
if (!getCompileInfo()
.equals(other.getCompileInfo())) return false;
}
if (hasDateStarted() != other.hasDateStarted()) return false;
if (hasDateStarted()) {
if (getDateStarted()
!= other.getDateStarted()) return false;
}
if (hasAdminAddress() != other.hasAdminAddress()) return false;
if (hasAdminAddress()) {
if (!getAdminAddress()
.equals(other.getAdminAddress())) return false;
}
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (hasDateCreated()) {
hash = (37 * hash) + DATECREATED_FIELD_NUMBER;
hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashLong(
getDateCreated());
}
if (hasDateModified()) {
hash = (37 * hash) + DATEMODIFIED_FIELD_NUMBER;
hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashLong(
getDateModified());
}
if (hasAddress()) {
hash = (37 * hash) + ADDRESS_FIELD_NUMBER;
hash = (53 * hash) + getAddress().hashCode();
}
if (hasStatus()) {
hash = (37 * hash) + STATUS_FIELD_NUMBER;
hash = (53 * hash) + getStatus().hashCode();
}
if (hasStateStoreVersion()) {
hash = (37 * hash) + STATESTOREVERSION_FIELD_NUMBER;
hash = (53 * hash) + getStateStoreVersion().hashCode();
}
if (hasVersion()) {
hash = (37 * hash) + VERSION_FIELD_NUMBER;
hash = (53 * hash) + getVersion().hashCode();
}
if (hasCompileInfo()) {
hash = (37 * hash) + COMPILEINFO_FIELD_NUMBER;
hash = (53 * hash) + getCompileInfo().hashCode();
}
if (hasDateStarted()) {
hash = (37 * hash) + DATESTARTED_FIELD_NUMBER;
hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashLong(
getDateStarted());
}
if (hasAdminAddress()) {
hash = (37 * hash) + ADMINADDRESS_FIELD_NUMBER;
hash = (53 * hash) + getAdminAddress().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.RouterRecordProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.RouterRecordProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RouterRecordProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RouterRecordProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
if (org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
getStateStoreVersionFieldBuilder();
}
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
dateCreated_ = 0L;
dateModified_ = 0L;
address_ = "";
status_ = "";
stateStoreVersion_ = null;
if (stateStoreVersionBuilder_ != null) {
stateStoreVersionBuilder_.dispose();
stateStoreVersionBuilder_ = null;
}
version_ = "";
compileInfo_ = "";
dateStarted_ = 0L;
adminAddress_ = "";
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RouterRecordProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto(this);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartial0(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto result) {
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) != 0)) {
result.dateCreated_ = dateCreated_;
to_bitField0_ |= 0x00000001;
}
if (((from_bitField0_ & 0x00000002) != 0)) {
result.dateModified_ = dateModified_;
to_bitField0_ |= 0x00000002;
}
if (((from_bitField0_ & 0x00000004) != 0)) {
result.address_ = address_;
to_bitField0_ |= 0x00000004;
}
if (((from_bitField0_ & 0x00000008) != 0)) {
result.status_ = status_;
to_bitField0_ |= 0x00000008;
}
if (((from_bitField0_ & 0x00000010) != 0)) {
result.stateStoreVersion_ = stateStoreVersionBuilder_ == null
? stateStoreVersion_
: stateStoreVersionBuilder_.build();
to_bitField0_ |= 0x00000010;
}
if (((from_bitField0_ & 0x00000020) != 0)) {
result.version_ = version_;
to_bitField0_ |= 0x00000020;
}
if (((from_bitField0_ & 0x00000040) != 0)) {
result.compileInfo_ = compileInfo_;
to_bitField0_ |= 0x00000040;
}
if (((from_bitField0_ & 0x00000080) != 0)) {
result.dateStarted_ = dateStarted_;
to_bitField0_ |= 0x00000080;
}
if (((from_bitField0_ & 0x00000100) != 0)) {
result.adminAddress_ = adminAddress_;
to_bitField0_ |= 0x00000100;
}
result.bitField0_ |= to_bitField0_;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto.getDefaultInstance()) return this;
if (other.hasDateCreated()) {
setDateCreated(other.getDateCreated());
}
if (other.hasDateModified()) {
setDateModified(other.getDateModified());
}
if (other.hasAddress()) {
address_ = other.address_;
bitField0_ |= 0x00000004;
onChanged();
}
if (other.hasStatus()) {
status_ = other.status_;
bitField0_ |= 0x00000008;
onChanged();
}
if (other.hasStateStoreVersion()) {
mergeStateStoreVersion(other.getStateStoreVersion());
}
if (other.hasVersion()) {
version_ = other.version_;
bitField0_ |= 0x00000020;
onChanged();
}
if (other.hasCompileInfo()) {
compileInfo_ = other.compileInfo_;
bitField0_ |= 0x00000040;
onChanged();
}
if (other.hasDateStarted()) {
setDateStarted(other.getDateStarted());
}
if (other.hasAdminAddress()) {
adminAddress_ = other.adminAddress_;
bitField0_ |= 0x00000100;
onChanged();
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 8: {
dateCreated_ = input.readUInt64();
bitField0_ |= 0x00000001;
break;
} // case 8
case 16: {
dateModified_ = input.readUInt64();
bitField0_ |= 0x00000002;
break;
} // case 16
case 26: {
address_ = input.readBytes();
bitField0_ |= 0x00000004;
break;
} // case 26
case 34: {
status_ = input.readBytes();
bitField0_ |= 0x00000008;
break;
} // case 34
case 42: {
input.readMessage(
getStateStoreVersionFieldBuilder().getBuilder(),
extensionRegistry);
bitField0_ |= 0x00000010;
break;
} // case 42
case 50: {
version_ = input.readBytes();
bitField0_ |= 0x00000020;
break;
} // case 50
case 58: {
compileInfo_ = input.readBytes();
bitField0_ |= 0x00000040;
break;
} // case 58
case 64: {
dateStarted_ = input.readUInt64();
bitField0_ |= 0x00000080;
break;
} // case 64
case 74: {
adminAddress_ = input.readBytes();
bitField0_ |= 0x00000100;
break;
} // case 74
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private long dateCreated_ ;
/**
* <code>optional uint64 dateCreated = 1;</code>
* @return Whether the dateCreated field is set.
*/
@java.lang.Override
public boolean hasDateCreated() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional uint64 dateCreated = 1;</code>
* @return The dateCreated.
*/
@java.lang.Override
public long getDateCreated() {
return dateCreated_;
}
/**
* <code>optional uint64 dateCreated = 1;</code>
* @param value The dateCreated to set.
* @return This builder for chaining.
*/
public Builder setDateCreated(long value) {
dateCreated_ = value;
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>optional uint64 dateCreated = 1;</code>
* @return This builder for chaining.
*/
public Builder clearDateCreated() {
bitField0_ = (bitField0_ & ~0x00000001);
dateCreated_ = 0L;
onChanged();
return this;
}
private long dateModified_ ;
/**
* <code>optional uint64 dateModified = 2;</code>
* @return Whether the dateModified field is set.
*/
@java.lang.Override
public boolean hasDateModified() {
return ((bitField0_ & 0x00000002) != 0);
}
/**
* <code>optional uint64 dateModified = 2;</code>
* @return The dateModified.
*/
@java.lang.Override
public long getDateModified() {
return dateModified_;
}
/**
* <code>optional uint64 dateModified = 2;</code>
* @param value The dateModified to set.
* @return This builder for chaining.
*/
public Builder setDateModified(long value) {
dateModified_ = value;
bitField0_ |= 0x00000002;
onChanged();
return this;
}
/**
* <code>optional uint64 dateModified = 2;</code>
* @return This builder for chaining.
*/
public Builder clearDateModified() {
bitField0_ = (bitField0_ & ~0x00000002);
dateModified_ = 0L;
onChanged();
return this;
}
private java.lang.Object address_ = "";
/**
* <code>optional string address = 3;</code>
* @return Whether the address field is set.
*/
public boolean hasAddress() {
return ((bitField0_ & 0x00000004) != 0);
}
/**
* <code>optional string address = 3;</code>
* @return The address.
*/
public java.lang.String getAddress() {
java.lang.Object ref = address_;
if (!(ref instanceof java.lang.String)) {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
address_ = s;
}
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* <code>optional string address = 3;</code>
* @return The bytes for address.
*/
public org.apache.hadoop.thirdparty.protobuf.ByteString
getAddressBytes() {
java.lang.Object ref = address_;
if (ref instanceof String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
address_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
/**
* <code>optional string address = 3;</code>
* @param value The address to set.
* @return This builder for chaining.
*/
public Builder setAddress(
java.lang.String value) {
if (value == null) { throw new NullPointerException(); }
address_ = value;
bitField0_ |= 0x00000004;
onChanged();
return this;
}
/**
* <code>optional string address = 3;</code>
* @return This builder for chaining.
*/
public Builder clearAddress() {
address_ = getDefaultInstance().getAddress();
bitField0_ = (bitField0_ & ~0x00000004);
onChanged();
return this;
}
/**
* <code>optional string address = 3;</code>
* @param value The bytes for address to set.
* @return This builder for chaining.
*/
public Builder setAddressBytes(
org.apache.hadoop.thirdparty.protobuf.ByteString value) {
if (value == null) { throw new NullPointerException(); }
address_ = value;
bitField0_ |= 0x00000004;
onChanged();
return this;
}
private java.lang.Object status_ = "";
/**
* <code>optional string status = 4;</code>
* @return Whether the status field is set.
*/
public boolean hasStatus() {
return ((bitField0_ & 0x00000008) != 0);
}
/**
* <code>optional string status = 4;</code>
* @return The status.
*/
public java.lang.String getStatus() {
java.lang.Object ref = status_;
if (!(ref instanceof java.lang.String)) {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
status_ = s;
}
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* <code>optional string status = 4;</code>
* @return The bytes for status.
*/
public org.apache.hadoop.thirdparty.protobuf.ByteString
getStatusBytes() {
java.lang.Object ref = status_;
if (ref instanceof String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
status_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
/**
* <code>optional string status = 4;</code>
* @param value The status to set.
* @return This builder for chaining.
*/
public Builder setStatus(
java.lang.String value) {
if (value == null) { throw new NullPointerException(); }
status_ = value;
bitField0_ |= 0x00000008;
onChanged();
return this;
}
/**
* <code>optional string status = 4;</code>
* @return This builder for chaining.
*/
public Builder clearStatus() {
status_ = getDefaultInstance().getStatus();
bitField0_ = (bitField0_ & ~0x00000008);
onChanged();
return this;
}
/**
* <code>optional string status = 4;</code>
* @param value The bytes for status to set.
* @return This builder for chaining.
*/
public Builder setStatusBytes(
org.apache.hadoop.thirdparty.protobuf.ByteString value) {
if (value == null) { throw new NullPointerException(); }
status_ = value;
bitField0_ |= 0x00000008;
onChanged();
return this;
}
private org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto stateStoreVersion_;
private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto.Builder, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProtoOrBuilder> stateStoreVersionBuilder_;
/**
* <code>optional .hadoop.hdfs.StateStoreVersionRecordProto stateStoreVersion = 5;</code>
* @return Whether the stateStoreVersion field is set.
*/
public boolean hasStateStoreVersion() {
return ((bitField0_ & 0x00000010) != 0);
}
/**
* <code>optional .hadoop.hdfs.StateStoreVersionRecordProto stateStoreVersion = 5;</code>
* @return The stateStoreVersion.
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto getStateStoreVersion() {
if (stateStoreVersionBuilder_ == null) {
return stateStoreVersion_ == null ? org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto.getDefaultInstance() : stateStoreVersion_;
} else {
return stateStoreVersionBuilder_.getMessage();
}
}
/**
* <code>optional .hadoop.hdfs.StateStoreVersionRecordProto stateStoreVersion = 5;</code>
*/
public Builder setStateStoreVersion(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto value) {
if (stateStoreVersionBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
stateStoreVersion_ = value;
} else {
stateStoreVersionBuilder_.setMessage(value);
}
bitField0_ |= 0x00000010;
onChanged();
return this;
}
/**
* <code>optional .hadoop.hdfs.StateStoreVersionRecordProto stateStoreVersion = 5;</code>
*/
public Builder setStateStoreVersion(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto.Builder builderForValue) {
if (stateStoreVersionBuilder_ == null) {
stateStoreVersion_ = builderForValue.build();
} else {
stateStoreVersionBuilder_.setMessage(builderForValue.build());
}
bitField0_ |= 0x00000010;
onChanged();
return this;
}
/**
* <code>optional .hadoop.hdfs.StateStoreVersionRecordProto stateStoreVersion = 5;</code>
*/
public Builder mergeStateStoreVersion(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto value) {
if (stateStoreVersionBuilder_ == null) {
if (((bitField0_ & 0x00000010) != 0) &&
stateStoreVersion_ != null &&
stateStoreVersion_ != org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto.getDefaultInstance()) {
getStateStoreVersionBuilder().mergeFrom(value);
} else {
stateStoreVersion_ = value;
}
} else {
stateStoreVersionBuilder_.mergeFrom(value);
}
if (stateStoreVersion_ != null) {
bitField0_ |= 0x00000010;
onChanged();
}
return this;
}
/**
* <code>optional .hadoop.hdfs.StateStoreVersionRecordProto stateStoreVersion = 5;</code>
*/
public Builder clearStateStoreVersion() {
bitField0_ = (bitField0_ & ~0x00000010);
stateStoreVersion_ = null;
if (stateStoreVersionBuilder_ != null) {
stateStoreVersionBuilder_.dispose();
stateStoreVersionBuilder_ = null;
}
onChanged();
return this;
}
/**
* <code>optional .hadoop.hdfs.StateStoreVersionRecordProto stateStoreVersion = 5;</code>
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto.Builder getStateStoreVersionBuilder() {
bitField0_ |= 0x00000010;
onChanged();
return getStateStoreVersionFieldBuilder().getBuilder();
}
/**
* <code>optional .hadoop.hdfs.StateStoreVersionRecordProto stateStoreVersion = 5;</code>
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProtoOrBuilder getStateStoreVersionOrBuilder() {
if (stateStoreVersionBuilder_ != null) {
return stateStoreVersionBuilder_.getMessageOrBuilder();
} else {
return stateStoreVersion_ == null ?
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto.getDefaultInstance() : stateStoreVersion_;
}
}
/**
* <code>optional .hadoop.hdfs.StateStoreVersionRecordProto stateStoreVersion = 5;</code>
*/
private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto.Builder, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProtoOrBuilder>
getStateStoreVersionFieldBuilder() {
if (stateStoreVersionBuilder_ == null) {
stateStoreVersionBuilder_ = new org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProto.Builder, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.StateStoreVersionRecordProtoOrBuilder>(
getStateStoreVersion(),
getParentForChildren(),
isClean());
stateStoreVersion_ = null;
}
return stateStoreVersionBuilder_;
}
private java.lang.Object version_ = "";
/**
* <code>optional string version = 6;</code>
* @return Whether the version field is set.
*/
public boolean hasVersion() {
return ((bitField0_ & 0x00000020) != 0);
}
/**
* <code>optional string version = 6;</code>
* @return The version.
*/
public java.lang.String getVersion() {
java.lang.Object ref = version_;
if (!(ref instanceof java.lang.String)) {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
version_ = s;
}
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* <code>optional string version = 6;</code>
* @return The bytes for version.
*/
public org.apache.hadoop.thirdparty.protobuf.ByteString
getVersionBytes() {
java.lang.Object ref = version_;
if (ref instanceof String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
version_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
/**
* <code>optional string version = 6;</code>
* @param value The version to set.
* @return This builder for chaining.
*/
public Builder setVersion(
java.lang.String value) {
if (value == null) { throw new NullPointerException(); }
version_ = value;
bitField0_ |= 0x00000020;
onChanged();
return this;
}
/**
* <code>optional string version = 6;</code>
* @return This builder for chaining.
*/
public Builder clearVersion() {
version_ = getDefaultInstance().getVersion();
bitField0_ = (bitField0_ & ~0x00000020);
onChanged();
return this;
}
/**
* <code>optional string version = 6;</code>
* @param value The bytes for version to set.
* @return This builder for chaining.
*/
public Builder setVersionBytes(
org.apache.hadoop.thirdparty.protobuf.ByteString value) {
if (value == null) { throw new NullPointerException(); }
version_ = value;
bitField0_ |= 0x00000020;
onChanged();
return this;
}
private java.lang.Object compileInfo_ = "";
/**
* <code>optional string compileInfo = 7;</code>
* @return Whether the compileInfo field is set.
*/
public boolean hasCompileInfo() {
return ((bitField0_ & 0x00000040) != 0);
}
/**
* <code>optional string compileInfo = 7;</code>
* @return The compileInfo.
*/
public java.lang.String getCompileInfo() {
java.lang.Object ref = compileInfo_;
if (!(ref instanceof java.lang.String)) {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
compileInfo_ = s;
}
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* <code>optional string compileInfo = 7;</code>
* @return The bytes for compileInfo.
*/
public org.apache.hadoop.thirdparty.protobuf.ByteString
getCompileInfoBytes() {
java.lang.Object ref = compileInfo_;
if (ref instanceof String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
compileInfo_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
/**
* <code>optional string compileInfo = 7;</code>
* @param value The compileInfo to set.
* @return This builder for chaining.
*/
public Builder setCompileInfo(
java.lang.String value) {
if (value == null) { throw new NullPointerException(); }
compileInfo_ = value;
bitField0_ |= 0x00000040;
onChanged();
return this;
}
/**
* <code>optional string compileInfo = 7;</code>
* @return This builder for chaining.
*/
public Builder clearCompileInfo() {
compileInfo_ = getDefaultInstance().getCompileInfo();
bitField0_ = (bitField0_ & ~0x00000040);
onChanged();
return this;
}
/**
* <code>optional string compileInfo = 7;</code>
* @param value The bytes for compileInfo to set.
* @return This builder for chaining.
*/
public Builder setCompileInfoBytes(
org.apache.hadoop.thirdparty.protobuf.ByteString value) {
if (value == null) { throw new NullPointerException(); }
compileInfo_ = value;
bitField0_ |= 0x00000040;
onChanged();
return this;
}
private long dateStarted_ ;
/**
* <code>optional uint64 dateStarted = 8;</code>
* @return Whether the dateStarted field is set.
*/
@java.lang.Override
public boolean hasDateStarted() {
return ((bitField0_ & 0x00000080) != 0);
}
/**
* <code>optional uint64 dateStarted = 8;</code>
* @return The dateStarted.
*/
@java.lang.Override
public long getDateStarted() {
return dateStarted_;
}
/**
* <code>optional uint64 dateStarted = 8;</code>
* @param value The dateStarted to set.
* @return This builder for chaining.
*/
public Builder setDateStarted(long value) {
dateStarted_ = value;
bitField0_ |= 0x00000080;
onChanged();
return this;
}
/**
* <code>optional uint64 dateStarted = 8;</code>
* @return This builder for chaining.
*/
public Builder clearDateStarted() {
bitField0_ = (bitField0_ & ~0x00000080);
dateStarted_ = 0L;
onChanged();
return this;
}
private java.lang.Object adminAddress_ = "";
/**
* <code>optional string adminAddress = 9;</code>
* @return Whether the adminAddress field is set.
*/
public boolean hasAdminAddress() {
return ((bitField0_ & 0x00000100) != 0);
}
/**
* <code>optional string adminAddress = 9;</code>
* @return The adminAddress.
*/
public java.lang.String getAdminAddress() {
java.lang.Object ref = adminAddress_;
if (!(ref instanceof java.lang.String)) {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
adminAddress_ = s;
}
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* <code>optional string adminAddress = 9;</code>
* @return The bytes for adminAddress.
*/
public org.apache.hadoop.thirdparty.protobuf.ByteString
getAdminAddressBytes() {
java.lang.Object ref = adminAddress_;
if (ref instanceof String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
adminAddress_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
/**
* <code>optional string adminAddress = 9;</code>
* @param value The adminAddress to set.
* @return This builder for chaining.
*/
public Builder setAdminAddress(
java.lang.String value) {
if (value == null) { throw new NullPointerException(); }
adminAddress_ = value;
bitField0_ |= 0x00000100;
onChanged();
return this;
}
/**
* <code>optional string adminAddress = 9;</code>
* @return This builder for chaining.
*/
public Builder clearAdminAddress() {
adminAddress_ = getDefaultInstance().getAdminAddress();
bitField0_ = (bitField0_ & ~0x00000100);
onChanged();
return this;
}
/**
* <code>optional string adminAddress = 9;</code>
* @param value The bytes for adminAddress to set.
* @return This builder for chaining.
*/
public Builder setAdminAddressBytes(
org.apache.hadoop.thirdparty.protobuf.ByteString value) {
if (value == null) { throw new NullPointerException(); }
adminAddress_ = value;
bitField0_ |= 0x00000100;
onChanged();
return this;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.RouterRecordProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.RouterRecordProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<RouterRecordProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<RouterRecordProto>() {
@java.lang.Override
public RouterRecordProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<RouterRecordProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<RouterRecordProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface GetRouterRegistrationRequestProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.GetRouterRegistrationRequestProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* <code>optional string routerId = 1;</code>
* @return Whether the routerId field is set.
*/
boolean hasRouterId();
/**
* <code>optional string routerId = 1;</code>
* @return The routerId.
*/
java.lang.String getRouterId();
/**
* <code>optional string routerId = 1;</code>
* @return The bytes for routerId.
*/
org.apache.hadoop.thirdparty.protobuf.ByteString
getRouterIdBytes();
}
/**
* Protobuf type {@code hadoop.hdfs.GetRouterRegistrationRequestProto}
*/
public static final class GetRouterRegistrationRequestProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.GetRouterRegistrationRequestProto)
GetRouterRegistrationRequestProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use GetRouterRegistrationRequestProto.newBuilder() to construct.
private GetRouterRegistrationRequestProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private GetRouterRegistrationRequestProto() {
routerId_ = "";
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new GetRouterRegistrationRequestProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetRouterRegistrationRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetRouterRegistrationRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationRequestProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationRequestProto.Builder.class);
}
private int bitField0_;
public static final int ROUTERID_FIELD_NUMBER = 1;
@SuppressWarnings("serial")
private volatile java.lang.Object routerId_ = "";
/**
* <code>optional string routerId = 1;</code>
* @return Whether the routerId field is set.
*/
@java.lang.Override
public boolean hasRouterId() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional string routerId = 1;</code>
* @return The routerId.
*/
@java.lang.Override
public java.lang.String getRouterId() {
java.lang.Object ref = routerId_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
routerId_ = s;
}
return s;
}
}
/**
* <code>optional string routerId = 1;</code>
* @return The bytes for routerId.
*/
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.ByteString
getRouterIdBytes() {
java.lang.Object ref = routerId_;
if (ref instanceof java.lang.String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
routerId_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) != 0)) {
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 1, routerId_);
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(1, routerId_);
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationRequestProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationRequestProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationRequestProto) obj;
if (hasRouterId() != other.hasRouterId()) return false;
if (hasRouterId()) {
if (!getRouterId()
.equals(other.getRouterId())) return false;
}
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (hasRouterId()) {
hash = (37 * hash) + ROUTERID_FIELD_NUMBER;
hash = (53 * hash) + getRouterId().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationRequestProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationRequestProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationRequestProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationRequestProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationRequestProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationRequestProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationRequestProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationRequestProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationRequestProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.GetRouterRegistrationRequestProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.GetRouterRegistrationRequestProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationRequestProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetRouterRegistrationRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetRouterRegistrationRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationRequestProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationRequestProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationRequestProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
routerId_ = "";
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetRouterRegistrationRequestProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationRequestProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationRequestProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationRequestProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationRequestProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationRequestProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationRequestProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationRequestProto(this);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartial0(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationRequestProto result) {
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) != 0)) {
result.routerId_ = routerId_;
to_bitField0_ |= 0x00000001;
}
result.bitField0_ |= to_bitField0_;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationRequestProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationRequestProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationRequestProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationRequestProto.getDefaultInstance()) return this;
if (other.hasRouterId()) {
routerId_ = other.routerId_;
bitField0_ |= 0x00000001;
onChanged();
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 10: {
routerId_ = input.readBytes();
bitField0_ |= 0x00000001;
break;
} // case 10
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private java.lang.Object routerId_ = "";
/**
* <code>optional string routerId = 1;</code>
* @return Whether the routerId field is set.
*/
public boolean hasRouterId() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional string routerId = 1;</code>
* @return The routerId.
*/
public java.lang.String getRouterId() {
java.lang.Object ref = routerId_;
if (!(ref instanceof java.lang.String)) {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
routerId_ = s;
}
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* <code>optional string routerId = 1;</code>
* @return The bytes for routerId.
*/
public org.apache.hadoop.thirdparty.protobuf.ByteString
getRouterIdBytes() {
java.lang.Object ref = routerId_;
if (ref instanceof String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
routerId_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
/**
* <code>optional string routerId = 1;</code>
* @param value The routerId to set.
* @return This builder for chaining.
*/
public Builder setRouterId(
java.lang.String value) {
if (value == null) { throw new NullPointerException(); }
routerId_ = value;
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>optional string routerId = 1;</code>
* @return This builder for chaining.
*/
public Builder clearRouterId() {
routerId_ = getDefaultInstance().getRouterId();
bitField0_ = (bitField0_ & ~0x00000001);
onChanged();
return this;
}
/**
* <code>optional string routerId = 1;</code>
* @param value The bytes for routerId to set.
* @return This builder for chaining.
*/
public Builder setRouterIdBytes(
org.apache.hadoop.thirdparty.protobuf.ByteString value) {
if (value == null) { throw new NullPointerException(); }
routerId_ = value;
bitField0_ |= 0x00000001;
onChanged();
return this;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.GetRouterRegistrationRequestProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.GetRouterRegistrationRequestProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationRequestProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationRequestProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationRequestProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<GetRouterRegistrationRequestProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<GetRouterRegistrationRequestProto>() {
@java.lang.Override
public GetRouterRegistrationRequestProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<GetRouterRegistrationRequestProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<GetRouterRegistrationRequestProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationRequestProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface GetRouterRegistrationResponseProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.GetRouterRegistrationResponseProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* <code>optional .hadoop.hdfs.RouterRecordProto router = 1;</code>
* @return Whether the router field is set.
*/
boolean hasRouter();
/**
* <code>optional .hadoop.hdfs.RouterRecordProto router = 1;</code>
* @return The router.
*/
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto getRouter();
/**
* <code>optional .hadoop.hdfs.RouterRecordProto router = 1;</code>
*/
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProtoOrBuilder getRouterOrBuilder();
}
/**
* Protobuf type {@code hadoop.hdfs.GetRouterRegistrationResponseProto}
*/
public static final class GetRouterRegistrationResponseProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.GetRouterRegistrationResponseProto)
GetRouterRegistrationResponseProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use GetRouterRegistrationResponseProto.newBuilder() to construct.
private GetRouterRegistrationResponseProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private GetRouterRegistrationResponseProto() {
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new GetRouterRegistrationResponseProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetRouterRegistrationResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetRouterRegistrationResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationResponseProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationResponseProto.Builder.class);
}
private int bitField0_;
public static final int ROUTER_FIELD_NUMBER = 1;
private org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto router_;
/**
* <code>optional .hadoop.hdfs.RouterRecordProto router = 1;</code>
* @return Whether the router field is set.
*/
@java.lang.Override
public boolean hasRouter() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional .hadoop.hdfs.RouterRecordProto router = 1;</code>
* @return The router.
*/
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto getRouter() {
return router_ == null ? org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto.getDefaultInstance() : router_;
}
/**
* <code>optional .hadoop.hdfs.RouterRecordProto router = 1;</code>
*/
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProtoOrBuilder getRouterOrBuilder() {
return router_ == null ? org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto.getDefaultInstance() : router_;
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) != 0)) {
output.writeMessage(1, getRouter());
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeMessageSize(1, getRouter());
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationResponseProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationResponseProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationResponseProto) obj;
if (hasRouter() != other.hasRouter()) return false;
if (hasRouter()) {
if (!getRouter()
.equals(other.getRouter())) return false;
}
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (hasRouter()) {
hash = (37 * hash) + ROUTER_FIELD_NUMBER;
hash = (53 * hash) + getRouter().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationResponseProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationResponseProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationResponseProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationResponseProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationResponseProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationResponseProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationResponseProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationResponseProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationResponseProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.GetRouterRegistrationResponseProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.GetRouterRegistrationResponseProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationResponseProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetRouterRegistrationResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetRouterRegistrationResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationResponseProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationResponseProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationResponseProto.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
if (org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
getRouterFieldBuilder();
}
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
router_ = null;
if (routerBuilder_ != null) {
routerBuilder_.dispose();
routerBuilder_ = null;
}
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetRouterRegistrationResponseProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationResponseProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationResponseProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationResponseProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationResponseProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationResponseProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationResponseProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationResponseProto(this);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartial0(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationResponseProto result) {
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) != 0)) {
result.router_ = routerBuilder_ == null
? router_
: routerBuilder_.build();
to_bitField0_ |= 0x00000001;
}
result.bitField0_ |= to_bitField0_;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationResponseProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationResponseProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationResponseProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationResponseProto.getDefaultInstance()) return this;
if (other.hasRouter()) {
mergeRouter(other.getRouter());
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 10: {
input.readMessage(
getRouterFieldBuilder().getBuilder(),
extensionRegistry);
bitField0_ |= 0x00000001;
break;
} // case 10
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto router_;
private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto.Builder, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProtoOrBuilder> routerBuilder_;
/**
* <code>optional .hadoop.hdfs.RouterRecordProto router = 1;</code>
* @return Whether the router field is set.
*/
public boolean hasRouter() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional .hadoop.hdfs.RouterRecordProto router = 1;</code>
* @return The router.
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto getRouter() {
if (routerBuilder_ == null) {
return router_ == null ? org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto.getDefaultInstance() : router_;
} else {
return routerBuilder_.getMessage();
}
}
/**
* <code>optional .hadoop.hdfs.RouterRecordProto router = 1;</code>
*/
public Builder setRouter(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto value) {
if (routerBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
router_ = value;
} else {
routerBuilder_.setMessage(value);
}
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>optional .hadoop.hdfs.RouterRecordProto router = 1;</code>
*/
public Builder setRouter(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto.Builder builderForValue) {
if (routerBuilder_ == null) {
router_ = builderForValue.build();
} else {
routerBuilder_.setMessage(builderForValue.build());
}
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>optional .hadoop.hdfs.RouterRecordProto router = 1;</code>
*/
public Builder mergeRouter(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto value) {
if (routerBuilder_ == null) {
if (((bitField0_ & 0x00000001) != 0) &&
router_ != null &&
router_ != org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto.getDefaultInstance()) {
getRouterBuilder().mergeFrom(value);
} else {
router_ = value;
}
} else {
routerBuilder_.mergeFrom(value);
}
if (router_ != null) {
bitField0_ |= 0x00000001;
onChanged();
}
return this;
}
/**
* <code>optional .hadoop.hdfs.RouterRecordProto router = 1;</code>
*/
public Builder clearRouter() {
bitField0_ = (bitField0_ & ~0x00000001);
router_ = null;
if (routerBuilder_ != null) {
routerBuilder_.dispose();
routerBuilder_ = null;
}
onChanged();
return this;
}
/**
* <code>optional .hadoop.hdfs.RouterRecordProto router = 1;</code>
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto.Builder getRouterBuilder() {
bitField0_ |= 0x00000001;
onChanged();
return getRouterFieldBuilder().getBuilder();
}
/**
* <code>optional .hadoop.hdfs.RouterRecordProto router = 1;</code>
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProtoOrBuilder getRouterOrBuilder() {
if (routerBuilder_ != null) {
return routerBuilder_.getMessageOrBuilder();
} else {
return router_ == null ?
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto.getDefaultInstance() : router_;
}
}
/**
* <code>optional .hadoop.hdfs.RouterRecordProto router = 1;</code>
*/
private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto.Builder, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProtoOrBuilder>
getRouterFieldBuilder() {
if (routerBuilder_ == null) {
routerBuilder_ = new org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto.Builder, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProtoOrBuilder>(
getRouter(),
getParentForChildren(),
isClean());
router_ = null;
}
return routerBuilder_;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.GetRouterRegistrationResponseProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.GetRouterRegistrationResponseProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationResponseProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationResponseProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationResponseProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<GetRouterRegistrationResponseProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<GetRouterRegistrationResponseProto>() {
@java.lang.Override
public GetRouterRegistrationResponseProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<GetRouterRegistrationResponseProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<GetRouterRegistrationResponseProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationResponseProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface GetRouterRegistrationsRequestProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.GetRouterRegistrationsRequestProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
}
/**
* Protobuf type {@code hadoop.hdfs.GetRouterRegistrationsRequestProto}
*/
public static final class GetRouterRegistrationsRequestProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.GetRouterRegistrationsRequestProto)
GetRouterRegistrationsRequestProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use GetRouterRegistrationsRequestProto.newBuilder() to construct.
private GetRouterRegistrationsRequestProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private GetRouterRegistrationsRequestProto() {
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new GetRouterRegistrationsRequestProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetRouterRegistrationsRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetRouterRegistrationsRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsRequestProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsRequestProto.Builder.class);
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsRequestProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsRequestProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsRequestProto) obj;
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsRequestProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsRequestProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsRequestProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsRequestProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsRequestProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsRequestProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsRequestProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsRequestProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsRequestProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.GetRouterRegistrationsRequestProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.GetRouterRegistrationsRequestProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsRequestProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetRouterRegistrationsRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetRouterRegistrationsRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsRequestProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsRequestProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsRequestProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetRouterRegistrationsRequestProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsRequestProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsRequestProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsRequestProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsRequestProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsRequestProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsRequestProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsRequestProto(this);
onBuilt();
return result;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsRequestProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsRequestProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsRequestProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsRequestProto.getDefaultInstance()) return this;
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.GetRouterRegistrationsRequestProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.GetRouterRegistrationsRequestProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsRequestProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsRequestProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsRequestProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<GetRouterRegistrationsRequestProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<GetRouterRegistrationsRequestProto>() {
@java.lang.Override
public GetRouterRegistrationsRequestProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<GetRouterRegistrationsRequestProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<GetRouterRegistrationsRequestProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsRequestProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface GetRouterRegistrationsResponseProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.GetRouterRegistrationsResponseProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* <code>optional uint64 timestamp = 1;</code>
* @return Whether the timestamp field is set.
*/
boolean hasTimestamp();
/**
* <code>optional uint64 timestamp = 1;</code>
* @return The timestamp.
*/
long getTimestamp();
/**
* <code>repeated .hadoop.hdfs.RouterRecordProto routers = 2;</code>
*/
java.util.List<org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto>
getRoutersList();
/**
* <code>repeated .hadoop.hdfs.RouterRecordProto routers = 2;</code>
*/
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto getRouters(int index);
/**
* <code>repeated .hadoop.hdfs.RouterRecordProto routers = 2;</code>
*/
int getRoutersCount();
/**
* <code>repeated .hadoop.hdfs.RouterRecordProto routers = 2;</code>
*/
java.util.List<? extends org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProtoOrBuilder>
getRoutersOrBuilderList();
/**
* <code>repeated .hadoop.hdfs.RouterRecordProto routers = 2;</code>
*/
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProtoOrBuilder getRoutersOrBuilder(
int index);
}
/**
* Protobuf type {@code hadoop.hdfs.GetRouterRegistrationsResponseProto}
*/
public static final class GetRouterRegistrationsResponseProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.GetRouterRegistrationsResponseProto)
GetRouterRegistrationsResponseProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use GetRouterRegistrationsResponseProto.newBuilder() to construct.
private GetRouterRegistrationsResponseProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private GetRouterRegistrationsResponseProto() {
routers_ = java.util.Collections.emptyList();
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new GetRouterRegistrationsResponseProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetRouterRegistrationsResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetRouterRegistrationsResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsResponseProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsResponseProto.Builder.class);
}
private int bitField0_;
public static final int TIMESTAMP_FIELD_NUMBER = 1;
private long timestamp_ = 0L;
/**
* <code>optional uint64 timestamp = 1;</code>
* @return Whether the timestamp field is set.
*/
@java.lang.Override
public boolean hasTimestamp() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional uint64 timestamp = 1;</code>
* @return The timestamp.
*/
@java.lang.Override
public long getTimestamp() {
return timestamp_;
}
public static final int ROUTERS_FIELD_NUMBER = 2;
@SuppressWarnings("serial")
private java.util.List<org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto> routers_;
/**
* <code>repeated .hadoop.hdfs.RouterRecordProto routers = 2;</code>
*/
@java.lang.Override
public java.util.List<org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto> getRoutersList() {
return routers_;
}
/**
* <code>repeated .hadoop.hdfs.RouterRecordProto routers = 2;</code>
*/
@java.lang.Override
public java.util.List<? extends org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProtoOrBuilder>
getRoutersOrBuilderList() {
return routers_;
}
/**
* <code>repeated .hadoop.hdfs.RouterRecordProto routers = 2;</code>
*/
@java.lang.Override
public int getRoutersCount() {
return routers_.size();
}
/**
* <code>repeated .hadoop.hdfs.RouterRecordProto routers = 2;</code>
*/
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto getRouters(int index) {
return routers_.get(index);
}
/**
* <code>repeated .hadoop.hdfs.RouterRecordProto routers = 2;</code>
*/
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProtoOrBuilder getRoutersOrBuilder(
int index) {
return routers_.get(index);
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) != 0)) {
output.writeUInt64(1, timestamp_);
}
for (int i = 0; i < routers_.size(); i++) {
output.writeMessage(2, routers_.get(i));
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeUInt64Size(1, timestamp_);
}
for (int i = 0; i < routers_.size(); i++) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeMessageSize(2, routers_.get(i));
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsResponseProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsResponseProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsResponseProto) obj;
if (hasTimestamp() != other.hasTimestamp()) return false;
if (hasTimestamp()) {
if (getTimestamp()
!= other.getTimestamp()) return false;
}
if (!getRoutersList()
.equals(other.getRoutersList())) return false;
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (hasTimestamp()) {
hash = (37 * hash) + TIMESTAMP_FIELD_NUMBER;
hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashLong(
getTimestamp());
}
if (getRoutersCount() > 0) {
hash = (37 * hash) + ROUTERS_FIELD_NUMBER;
hash = (53 * hash) + getRoutersList().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsResponseProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsResponseProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsResponseProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsResponseProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsResponseProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsResponseProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsResponseProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsResponseProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsResponseProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.GetRouterRegistrationsResponseProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.GetRouterRegistrationsResponseProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsResponseProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetRouterRegistrationsResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetRouterRegistrationsResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsResponseProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsResponseProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsResponseProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
timestamp_ = 0L;
if (routersBuilder_ == null) {
routers_ = java.util.Collections.emptyList();
} else {
routers_ = null;
routersBuilder_.clear();
}
bitField0_ = (bitField0_ & ~0x00000002);
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetRouterRegistrationsResponseProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsResponseProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsResponseProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsResponseProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsResponseProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsResponseProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsResponseProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsResponseProto(this);
buildPartialRepeatedFields(result);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartialRepeatedFields(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsResponseProto result) {
if (routersBuilder_ == null) {
if (((bitField0_ & 0x00000002) != 0)) {
routers_ = java.util.Collections.unmodifiableList(routers_);
bitField0_ = (bitField0_ & ~0x00000002);
}
result.routers_ = routers_;
} else {
result.routers_ = routersBuilder_.build();
}
}
private void buildPartial0(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsResponseProto result) {
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) != 0)) {
result.timestamp_ = timestamp_;
to_bitField0_ |= 0x00000001;
}
result.bitField0_ |= to_bitField0_;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsResponseProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsResponseProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsResponseProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsResponseProto.getDefaultInstance()) return this;
if (other.hasTimestamp()) {
setTimestamp(other.getTimestamp());
}
if (routersBuilder_ == null) {
if (!other.routers_.isEmpty()) {
if (routers_.isEmpty()) {
routers_ = other.routers_;
bitField0_ = (bitField0_ & ~0x00000002);
} else {
ensureRoutersIsMutable();
routers_.addAll(other.routers_);
}
onChanged();
}
} else {
if (!other.routers_.isEmpty()) {
if (routersBuilder_.isEmpty()) {
routersBuilder_.dispose();
routersBuilder_ = null;
routers_ = other.routers_;
bitField0_ = (bitField0_ & ~0x00000002);
routersBuilder_ =
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
getRoutersFieldBuilder() : null;
} else {
routersBuilder_.addAllMessages(other.routers_);
}
}
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 8: {
timestamp_ = input.readUInt64();
bitField0_ |= 0x00000001;
break;
} // case 8
case 18: {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto m =
input.readMessage(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto.PARSER,
extensionRegistry);
if (routersBuilder_ == null) {
ensureRoutersIsMutable();
routers_.add(m);
} else {
routersBuilder_.addMessage(m);
}
break;
} // case 18
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private long timestamp_ ;
/**
* <code>optional uint64 timestamp = 1;</code>
* @return Whether the timestamp field is set.
*/
@java.lang.Override
public boolean hasTimestamp() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional uint64 timestamp = 1;</code>
* @return The timestamp.
*/
@java.lang.Override
public long getTimestamp() {
return timestamp_;
}
/**
* <code>optional uint64 timestamp = 1;</code>
* @param value The timestamp to set.
* @return This builder for chaining.
*/
public Builder setTimestamp(long value) {
timestamp_ = value;
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>optional uint64 timestamp = 1;</code>
* @return This builder for chaining.
*/
public Builder clearTimestamp() {
bitField0_ = (bitField0_ & ~0x00000001);
timestamp_ = 0L;
onChanged();
return this;
}
private java.util.List<org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto> routers_ =
java.util.Collections.emptyList();
private void ensureRoutersIsMutable() {
if (!((bitField0_ & 0x00000002) != 0)) {
routers_ = new java.util.ArrayList<org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto>(routers_);
bitField0_ |= 0x00000002;
}
}
private org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto.Builder, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProtoOrBuilder> routersBuilder_;
/**
* <code>repeated .hadoop.hdfs.RouterRecordProto routers = 2;</code>
*/
public java.util.List<org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto> getRoutersList() {
if (routersBuilder_ == null) {
return java.util.Collections.unmodifiableList(routers_);
} else {
return routersBuilder_.getMessageList();
}
}
/**
* <code>repeated .hadoop.hdfs.RouterRecordProto routers = 2;</code>
*/
public int getRoutersCount() {
if (routersBuilder_ == null) {
return routers_.size();
} else {
return routersBuilder_.getCount();
}
}
/**
* <code>repeated .hadoop.hdfs.RouterRecordProto routers = 2;</code>
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto getRouters(int index) {
if (routersBuilder_ == null) {
return routers_.get(index);
} else {
return routersBuilder_.getMessage(index);
}
}
/**
* <code>repeated .hadoop.hdfs.RouterRecordProto routers = 2;</code>
*/
public Builder setRouters(
int index, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto value) {
if (routersBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
ensureRoutersIsMutable();
routers_.set(index, value);
onChanged();
} else {
routersBuilder_.setMessage(index, value);
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.RouterRecordProto routers = 2;</code>
*/
public Builder setRouters(
int index, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto.Builder builderForValue) {
if (routersBuilder_ == null) {
ensureRoutersIsMutable();
routers_.set(index, builderForValue.build());
onChanged();
} else {
routersBuilder_.setMessage(index, builderForValue.build());
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.RouterRecordProto routers = 2;</code>
*/
public Builder addRouters(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto value) {
if (routersBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
ensureRoutersIsMutable();
routers_.add(value);
onChanged();
} else {
routersBuilder_.addMessage(value);
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.RouterRecordProto routers = 2;</code>
*/
public Builder addRouters(
int index, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto value) {
if (routersBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
ensureRoutersIsMutable();
routers_.add(index, value);
onChanged();
} else {
routersBuilder_.addMessage(index, value);
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.RouterRecordProto routers = 2;</code>
*/
public Builder addRouters(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto.Builder builderForValue) {
if (routersBuilder_ == null) {
ensureRoutersIsMutable();
routers_.add(builderForValue.build());
onChanged();
} else {
routersBuilder_.addMessage(builderForValue.build());
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.RouterRecordProto routers = 2;</code>
*/
public Builder addRouters(
int index, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto.Builder builderForValue) {
if (routersBuilder_ == null) {
ensureRoutersIsMutable();
routers_.add(index, builderForValue.build());
onChanged();
} else {
routersBuilder_.addMessage(index, builderForValue.build());
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.RouterRecordProto routers = 2;</code>
*/
public Builder addAllRouters(
java.lang.Iterable<? extends org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto> values) {
if (routersBuilder_ == null) {
ensureRoutersIsMutable();
org.apache.hadoop.thirdparty.protobuf.AbstractMessageLite.Builder.addAll(
values, routers_);
onChanged();
} else {
routersBuilder_.addAllMessages(values);
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.RouterRecordProto routers = 2;</code>
*/
public Builder clearRouters() {
if (routersBuilder_ == null) {
routers_ = java.util.Collections.emptyList();
bitField0_ = (bitField0_ & ~0x00000002);
onChanged();
} else {
routersBuilder_.clear();
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.RouterRecordProto routers = 2;</code>
*/
public Builder removeRouters(int index) {
if (routersBuilder_ == null) {
ensureRoutersIsMutable();
routers_.remove(index);
onChanged();
} else {
routersBuilder_.remove(index);
}
return this;
}
/**
* <code>repeated .hadoop.hdfs.RouterRecordProto routers = 2;</code>
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto.Builder getRoutersBuilder(
int index) {
return getRoutersFieldBuilder().getBuilder(index);
}
/**
* <code>repeated .hadoop.hdfs.RouterRecordProto routers = 2;</code>
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProtoOrBuilder getRoutersOrBuilder(
int index) {
if (routersBuilder_ == null) {
return routers_.get(index); } else {
return routersBuilder_.getMessageOrBuilder(index);
}
}
/**
* <code>repeated .hadoop.hdfs.RouterRecordProto routers = 2;</code>
*/
public java.util.List<? extends org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProtoOrBuilder>
getRoutersOrBuilderList() {
if (routersBuilder_ != null) {
return routersBuilder_.getMessageOrBuilderList();
} else {
return java.util.Collections.unmodifiableList(routers_);
}
}
/**
* <code>repeated .hadoop.hdfs.RouterRecordProto routers = 2;</code>
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto.Builder addRoutersBuilder() {
return getRoutersFieldBuilder().addBuilder(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto.getDefaultInstance());
}
/**
* <code>repeated .hadoop.hdfs.RouterRecordProto routers = 2;</code>
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto.Builder addRoutersBuilder(
int index) {
return getRoutersFieldBuilder().addBuilder(
index, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto.getDefaultInstance());
}
/**
* <code>repeated .hadoop.hdfs.RouterRecordProto routers = 2;</code>
*/
public java.util.List<org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto.Builder>
getRoutersBuilderList() {
return getRoutersFieldBuilder().getBuilderList();
}
private org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto.Builder, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProtoOrBuilder>
getRoutersFieldBuilder() {
if (routersBuilder_ == null) {
routersBuilder_ = new org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto.Builder, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProtoOrBuilder>(
routers_,
((bitField0_ & 0x00000002) != 0),
getParentForChildren(),
isClean());
routers_ = null;
}
return routersBuilder_;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.GetRouterRegistrationsResponseProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.GetRouterRegistrationsResponseProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsResponseProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsResponseProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsResponseProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<GetRouterRegistrationsResponseProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<GetRouterRegistrationsResponseProto>() {
@java.lang.Override
public GetRouterRegistrationsResponseProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<GetRouterRegistrationsResponseProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<GetRouterRegistrationsResponseProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetRouterRegistrationsResponseProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface RouterHeartbeatRequestProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.RouterHeartbeatRequestProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* <code>optional .hadoop.hdfs.RouterRecordProto router = 1;</code>
* @return Whether the router field is set.
*/
boolean hasRouter();
/**
* <code>optional .hadoop.hdfs.RouterRecordProto router = 1;</code>
* @return The router.
*/
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto getRouter();
/**
* <code>optional .hadoop.hdfs.RouterRecordProto router = 1;</code>
*/
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProtoOrBuilder getRouterOrBuilder();
}
/**
* Protobuf type {@code hadoop.hdfs.RouterHeartbeatRequestProto}
*/
public static final class RouterHeartbeatRequestProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.RouterHeartbeatRequestProto)
RouterHeartbeatRequestProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use RouterHeartbeatRequestProto.newBuilder() to construct.
private RouterHeartbeatRequestProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private RouterHeartbeatRequestProto() {
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new RouterHeartbeatRequestProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RouterHeartbeatRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RouterHeartbeatRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatRequestProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatRequestProto.Builder.class);
}
private int bitField0_;
public static final int ROUTER_FIELD_NUMBER = 1;
private org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto router_;
/**
* <code>optional .hadoop.hdfs.RouterRecordProto router = 1;</code>
* @return Whether the router field is set.
*/
@java.lang.Override
public boolean hasRouter() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional .hadoop.hdfs.RouterRecordProto router = 1;</code>
* @return The router.
*/
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto getRouter() {
return router_ == null ? org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto.getDefaultInstance() : router_;
}
/**
* <code>optional .hadoop.hdfs.RouterRecordProto router = 1;</code>
*/
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProtoOrBuilder getRouterOrBuilder() {
return router_ == null ? org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto.getDefaultInstance() : router_;
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) != 0)) {
output.writeMessage(1, getRouter());
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeMessageSize(1, getRouter());
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatRequestProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatRequestProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatRequestProto) obj;
if (hasRouter() != other.hasRouter()) return false;
if (hasRouter()) {
if (!getRouter()
.equals(other.getRouter())) return false;
}
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (hasRouter()) {
hash = (37 * hash) + ROUTER_FIELD_NUMBER;
hash = (53 * hash) + getRouter().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatRequestProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatRequestProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatRequestProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatRequestProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatRequestProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatRequestProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatRequestProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatRequestProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatRequestProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.RouterHeartbeatRequestProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.RouterHeartbeatRequestProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatRequestProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RouterHeartbeatRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RouterHeartbeatRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatRequestProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatRequestProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatRequestProto.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
if (org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
getRouterFieldBuilder();
}
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
router_ = null;
if (routerBuilder_ != null) {
routerBuilder_.dispose();
routerBuilder_ = null;
}
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RouterHeartbeatRequestProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatRequestProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatRequestProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatRequestProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatRequestProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatRequestProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatRequestProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatRequestProto(this);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartial0(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatRequestProto result) {
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) != 0)) {
result.router_ = routerBuilder_ == null
? router_
: routerBuilder_.build();
to_bitField0_ |= 0x00000001;
}
result.bitField0_ |= to_bitField0_;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatRequestProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatRequestProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatRequestProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatRequestProto.getDefaultInstance()) return this;
if (other.hasRouter()) {
mergeRouter(other.getRouter());
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 10: {
input.readMessage(
getRouterFieldBuilder().getBuilder(),
extensionRegistry);
bitField0_ |= 0x00000001;
break;
} // case 10
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto router_;
private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto.Builder, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProtoOrBuilder> routerBuilder_;
/**
* <code>optional .hadoop.hdfs.RouterRecordProto router = 1;</code>
* @return Whether the router field is set.
*/
public boolean hasRouter() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional .hadoop.hdfs.RouterRecordProto router = 1;</code>
* @return The router.
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto getRouter() {
if (routerBuilder_ == null) {
return router_ == null ? org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto.getDefaultInstance() : router_;
} else {
return routerBuilder_.getMessage();
}
}
/**
* <code>optional .hadoop.hdfs.RouterRecordProto router = 1;</code>
*/
public Builder setRouter(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto value) {
if (routerBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
router_ = value;
} else {
routerBuilder_.setMessage(value);
}
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>optional .hadoop.hdfs.RouterRecordProto router = 1;</code>
*/
public Builder setRouter(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto.Builder builderForValue) {
if (routerBuilder_ == null) {
router_ = builderForValue.build();
} else {
routerBuilder_.setMessage(builderForValue.build());
}
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>optional .hadoop.hdfs.RouterRecordProto router = 1;</code>
*/
public Builder mergeRouter(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto value) {
if (routerBuilder_ == null) {
if (((bitField0_ & 0x00000001) != 0) &&
router_ != null &&
router_ != org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto.getDefaultInstance()) {
getRouterBuilder().mergeFrom(value);
} else {
router_ = value;
}
} else {
routerBuilder_.mergeFrom(value);
}
if (router_ != null) {
bitField0_ |= 0x00000001;
onChanged();
}
return this;
}
/**
* <code>optional .hadoop.hdfs.RouterRecordProto router = 1;</code>
*/
public Builder clearRouter() {
bitField0_ = (bitField0_ & ~0x00000001);
router_ = null;
if (routerBuilder_ != null) {
routerBuilder_.dispose();
routerBuilder_ = null;
}
onChanged();
return this;
}
/**
* <code>optional .hadoop.hdfs.RouterRecordProto router = 1;</code>
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto.Builder getRouterBuilder() {
bitField0_ |= 0x00000001;
onChanged();
return getRouterFieldBuilder().getBuilder();
}
/**
* <code>optional .hadoop.hdfs.RouterRecordProto router = 1;</code>
*/
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProtoOrBuilder getRouterOrBuilder() {
if (routerBuilder_ != null) {
return routerBuilder_.getMessageOrBuilder();
} else {
return router_ == null ?
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto.getDefaultInstance() : router_;
}
}
/**
* <code>optional .hadoop.hdfs.RouterRecordProto router = 1;</code>
*/
private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto.Builder, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProtoOrBuilder>
getRouterFieldBuilder() {
if (routerBuilder_ == null) {
routerBuilder_ = new org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProto.Builder, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterRecordProtoOrBuilder>(
getRouter(),
getParentForChildren(),
isClean());
router_ = null;
}
return routerBuilder_;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.RouterHeartbeatRequestProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.RouterHeartbeatRequestProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatRequestProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatRequestProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatRequestProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<RouterHeartbeatRequestProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<RouterHeartbeatRequestProto>() {
@java.lang.Override
public RouterHeartbeatRequestProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<RouterHeartbeatRequestProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<RouterHeartbeatRequestProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatRequestProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface RouterHeartbeatResponseProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.RouterHeartbeatResponseProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* <code>optional bool status = 1;</code>
* @return Whether the status field is set.
*/
boolean hasStatus();
/**
* <code>optional bool status = 1;</code>
* @return The status.
*/
boolean getStatus();
}
/**
* Protobuf type {@code hadoop.hdfs.RouterHeartbeatResponseProto}
*/
public static final class RouterHeartbeatResponseProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.RouterHeartbeatResponseProto)
RouterHeartbeatResponseProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use RouterHeartbeatResponseProto.newBuilder() to construct.
private RouterHeartbeatResponseProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private RouterHeartbeatResponseProto() {
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new RouterHeartbeatResponseProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RouterHeartbeatResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RouterHeartbeatResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatResponseProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatResponseProto.Builder.class);
}
private int bitField0_;
public static final int STATUS_FIELD_NUMBER = 1;
private boolean status_ = false;
/**
* <code>optional bool status = 1;</code>
* @return Whether the status field is set.
*/
@java.lang.Override
public boolean hasStatus() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional bool status = 1;</code>
* @return The status.
*/
@java.lang.Override
public boolean getStatus() {
return status_;
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) != 0)) {
output.writeBool(1, status_);
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeBoolSize(1, status_);
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatResponseProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatResponseProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatResponseProto) obj;
if (hasStatus() != other.hasStatus()) return false;
if (hasStatus()) {
if (getStatus()
!= other.getStatus()) return false;
}
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (hasStatus()) {
hash = (37 * hash) + STATUS_FIELD_NUMBER;
hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashBoolean(
getStatus());
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatResponseProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatResponseProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatResponseProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatResponseProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatResponseProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatResponseProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatResponseProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatResponseProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatResponseProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.RouterHeartbeatResponseProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.RouterHeartbeatResponseProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatResponseProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RouterHeartbeatResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RouterHeartbeatResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatResponseProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatResponseProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatResponseProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
status_ = false;
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RouterHeartbeatResponseProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatResponseProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatResponseProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatResponseProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatResponseProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatResponseProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatResponseProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatResponseProto(this);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartial0(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatResponseProto result) {
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) != 0)) {
result.status_ = status_;
to_bitField0_ |= 0x00000001;
}
result.bitField0_ |= to_bitField0_;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatResponseProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatResponseProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatResponseProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatResponseProto.getDefaultInstance()) return this;
if (other.hasStatus()) {
setStatus(other.getStatus());
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 8: {
status_ = input.readBool();
bitField0_ |= 0x00000001;
break;
} // case 8
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private boolean status_ ;
/**
* <code>optional bool status = 1;</code>
* @return Whether the status field is set.
*/
@java.lang.Override
public boolean hasStatus() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional bool status = 1;</code>
* @return The status.
*/
@java.lang.Override
public boolean getStatus() {
return status_;
}
/**
* <code>optional bool status = 1;</code>
* @param value The status to set.
* @return This builder for chaining.
*/
public Builder setStatus(boolean value) {
status_ = value;
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>optional bool status = 1;</code>
* @return This builder for chaining.
*/
public Builder clearStatus() {
bitField0_ = (bitField0_ & ~0x00000001);
status_ = false;
onChanged();
return this;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.RouterHeartbeatResponseProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.RouterHeartbeatResponseProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatResponseProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatResponseProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatResponseProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<RouterHeartbeatResponseProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<RouterHeartbeatResponseProto>() {
@java.lang.Override
public RouterHeartbeatResponseProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<RouterHeartbeatResponseProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<RouterHeartbeatResponseProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RouterHeartbeatResponseProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface RefreshMountTableEntriesRequestProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.RefreshMountTableEntriesRequestProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
}
/**
* Protobuf type {@code hadoop.hdfs.RefreshMountTableEntriesRequestProto}
*/
public static final class RefreshMountTableEntriesRequestProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.RefreshMountTableEntriesRequestProto)
RefreshMountTableEntriesRequestProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use RefreshMountTableEntriesRequestProto.newBuilder() to construct.
private RefreshMountTableEntriesRequestProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private RefreshMountTableEntriesRequestProto() {
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new RefreshMountTableEntriesRequestProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RefreshMountTableEntriesRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RefreshMountTableEntriesRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesRequestProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesRequestProto.Builder.class);
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesRequestProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesRequestProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesRequestProto) obj;
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesRequestProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesRequestProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesRequestProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesRequestProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesRequestProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesRequestProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesRequestProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesRequestProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesRequestProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.RefreshMountTableEntriesRequestProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.RefreshMountTableEntriesRequestProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesRequestProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RefreshMountTableEntriesRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RefreshMountTableEntriesRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesRequestProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesRequestProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesRequestProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RefreshMountTableEntriesRequestProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesRequestProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesRequestProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesRequestProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesRequestProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesRequestProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesRequestProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesRequestProto(this);
onBuilt();
return result;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesRequestProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesRequestProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesRequestProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesRequestProto.getDefaultInstance()) return this;
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.RefreshMountTableEntriesRequestProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.RefreshMountTableEntriesRequestProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesRequestProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesRequestProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesRequestProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<RefreshMountTableEntriesRequestProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<RefreshMountTableEntriesRequestProto>() {
@java.lang.Override
public RefreshMountTableEntriesRequestProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<RefreshMountTableEntriesRequestProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<RefreshMountTableEntriesRequestProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesRequestProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface RefreshMountTableEntriesResponseProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.RefreshMountTableEntriesResponseProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* <code>optional bool result = 1;</code>
* @return Whether the result field is set.
*/
boolean hasResult();
/**
* <code>optional bool result = 1;</code>
* @return The result.
*/
boolean getResult();
}
/**
* Protobuf type {@code hadoop.hdfs.RefreshMountTableEntriesResponseProto}
*/
public static final class RefreshMountTableEntriesResponseProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.RefreshMountTableEntriesResponseProto)
RefreshMountTableEntriesResponseProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use RefreshMountTableEntriesResponseProto.newBuilder() to construct.
private RefreshMountTableEntriesResponseProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private RefreshMountTableEntriesResponseProto() {
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new RefreshMountTableEntriesResponseProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RefreshMountTableEntriesResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RefreshMountTableEntriesResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesResponseProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesResponseProto.Builder.class);
}
private int bitField0_;
public static final int RESULT_FIELD_NUMBER = 1;
private boolean result_ = false;
/**
* <code>optional bool result = 1;</code>
* @return Whether the result field is set.
*/
@java.lang.Override
public boolean hasResult() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional bool result = 1;</code>
* @return The result.
*/
@java.lang.Override
public boolean getResult() {
return result_;
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) != 0)) {
output.writeBool(1, result_);
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeBoolSize(1, result_);
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesResponseProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesResponseProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesResponseProto) obj;
if (hasResult() != other.hasResult()) return false;
if (hasResult()) {
if (getResult()
!= other.getResult()) return false;
}
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (hasResult()) {
hash = (37 * hash) + RESULT_FIELD_NUMBER;
hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashBoolean(
getResult());
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesResponseProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesResponseProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesResponseProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesResponseProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesResponseProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesResponseProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesResponseProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesResponseProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesResponseProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.RefreshMountTableEntriesResponseProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.RefreshMountTableEntriesResponseProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesResponseProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RefreshMountTableEntriesResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RefreshMountTableEntriesResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesResponseProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesResponseProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesResponseProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
result_ = false;
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RefreshMountTableEntriesResponseProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesResponseProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesResponseProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesResponseProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesResponseProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesResponseProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesResponseProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesResponseProto(this);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartial0(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesResponseProto result) {
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) != 0)) {
result.result_ = result_;
to_bitField0_ |= 0x00000001;
}
result.bitField0_ |= to_bitField0_;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesResponseProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesResponseProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesResponseProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesResponseProto.getDefaultInstance()) return this;
if (other.hasResult()) {
setResult(other.getResult());
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 8: {
result_ = input.readBool();
bitField0_ |= 0x00000001;
break;
} // case 8
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private boolean result_ ;
/**
* <code>optional bool result = 1;</code>
* @return Whether the result field is set.
*/
@java.lang.Override
public boolean hasResult() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional bool result = 1;</code>
* @return The result.
*/
@java.lang.Override
public boolean getResult() {
return result_;
}
/**
* <code>optional bool result = 1;</code>
* @param value The result to set.
* @return This builder for chaining.
*/
public Builder setResult(boolean value) {
result_ = value;
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>optional bool result = 1;</code>
* @return This builder for chaining.
*/
public Builder clearResult() {
bitField0_ = (bitField0_ & ~0x00000001);
result_ = false;
onChanged();
return this;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.RefreshMountTableEntriesResponseProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.RefreshMountTableEntriesResponseProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesResponseProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesResponseProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesResponseProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<RefreshMountTableEntriesResponseProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<RefreshMountTableEntriesResponseProto>() {
@java.lang.Override
public RefreshMountTableEntriesResponseProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<RefreshMountTableEntriesResponseProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<RefreshMountTableEntriesResponseProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshMountTableEntriesResponseProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface RefreshSuperUserGroupsConfigurationRequestProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.RefreshSuperUserGroupsConfigurationRequestProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
}
/**
* Protobuf type {@code hadoop.hdfs.RefreshSuperUserGroupsConfigurationRequestProto}
*/
public static final class RefreshSuperUserGroupsConfigurationRequestProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.RefreshSuperUserGroupsConfigurationRequestProto)
RefreshSuperUserGroupsConfigurationRequestProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use RefreshSuperUserGroupsConfigurationRequestProto.newBuilder() to construct.
private RefreshSuperUserGroupsConfigurationRequestProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private RefreshSuperUserGroupsConfigurationRequestProto() {
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new RefreshSuperUserGroupsConfigurationRequestProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RefreshSuperUserGroupsConfigurationRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RefreshSuperUserGroupsConfigurationRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationRequestProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationRequestProto.Builder.class);
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationRequestProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationRequestProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationRequestProto) obj;
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationRequestProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationRequestProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationRequestProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationRequestProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationRequestProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationRequestProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationRequestProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationRequestProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationRequestProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.RefreshSuperUserGroupsConfigurationRequestProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.RefreshSuperUserGroupsConfigurationRequestProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationRequestProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RefreshSuperUserGroupsConfigurationRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RefreshSuperUserGroupsConfigurationRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationRequestProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationRequestProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationRequestProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RefreshSuperUserGroupsConfigurationRequestProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationRequestProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationRequestProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationRequestProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationRequestProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationRequestProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationRequestProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationRequestProto(this);
onBuilt();
return result;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationRequestProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationRequestProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationRequestProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationRequestProto.getDefaultInstance()) return this;
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.RefreshSuperUserGroupsConfigurationRequestProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.RefreshSuperUserGroupsConfigurationRequestProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationRequestProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationRequestProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationRequestProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<RefreshSuperUserGroupsConfigurationRequestProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<RefreshSuperUserGroupsConfigurationRequestProto>() {
@java.lang.Override
public RefreshSuperUserGroupsConfigurationRequestProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<RefreshSuperUserGroupsConfigurationRequestProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<RefreshSuperUserGroupsConfigurationRequestProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationRequestProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface RefreshSuperUserGroupsConfigurationResponseProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.RefreshSuperUserGroupsConfigurationResponseProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* <code>optional bool status = 1;</code>
* @return Whether the status field is set.
*/
boolean hasStatus();
/**
* <code>optional bool status = 1;</code>
* @return The status.
*/
boolean getStatus();
}
/**
* Protobuf type {@code hadoop.hdfs.RefreshSuperUserGroupsConfigurationResponseProto}
*/
public static final class RefreshSuperUserGroupsConfigurationResponseProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.RefreshSuperUserGroupsConfigurationResponseProto)
RefreshSuperUserGroupsConfigurationResponseProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use RefreshSuperUserGroupsConfigurationResponseProto.newBuilder() to construct.
private RefreshSuperUserGroupsConfigurationResponseProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private RefreshSuperUserGroupsConfigurationResponseProto() {
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new RefreshSuperUserGroupsConfigurationResponseProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RefreshSuperUserGroupsConfigurationResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RefreshSuperUserGroupsConfigurationResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationResponseProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationResponseProto.Builder.class);
}
private int bitField0_;
public static final int STATUS_FIELD_NUMBER = 1;
private boolean status_ = false;
/**
* <code>optional bool status = 1;</code>
* @return Whether the status field is set.
*/
@java.lang.Override
public boolean hasStatus() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional bool status = 1;</code>
* @return The status.
*/
@java.lang.Override
public boolean getStatus() {
return status_;
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) != 0)) {
output.writeBool(1, status_);
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeBoolSize(1, status_);
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationResponseProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationResponseProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationResponseProto) obj;
if (hasStatus() != other.hasStatus()) return false;
if (hasStatus()) {
if (getStatus()
!= other.getStatus()) return false;
}
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (hasStatus()) {
hash = (37 * hash) + STATUS_FIELD_NUMBER;
hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashBoolean(
getStatus());
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationResponseProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationResponseProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationResponseProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationResponseProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationResponseProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationResponseProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationResponseProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationResponseProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationResponseProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.RefreshSuperUserGroupsConfigurationResponseProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.RefreshSuperUserGroupsConfigurationResponseProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationResponseProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RefreshSuperUserGroupsConfigurationResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RefreshSuperUserGroupsConfigurationResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationResponseProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationResponseProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationResponseProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
status_ = false;
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_RefreshSuperUserGroupsConfigurationResponseProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationResponseProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationResponseProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationResponseProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationResponseProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationResponseProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationResponseProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationResponseProto(this);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartial0(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationResponseProto result) {
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) != 0)) {
result.status_ = status_;
to_bitField0_ |= 0x00000001;
}
result.bitField0_ |= to_bitField0_;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationResponseProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationResponseProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationResponseProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationResponseProto.getDefaultInstance()) return this;
if (other.hasStatus()) {
setStatus(other.getStatus());
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 8: {
status_ = input.readBool();
bitField0_ |= 0x00000001;
break;
} // case 8
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private boolean status_ ;
/**
* <code>optional bool status = 1;</code>
* @return Whether the status field is set.
*/
@java.lang.Override
public boolean hasStatus() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional bool status = 1;</code>
* @return The status.
*/
@java.lang.Override
public boolean getStatus() {
return status_;
}
/**
* <code>optional bool status = 1;</code>
* @param value The status to set.
* @return This builder for chaining.
*/
public Builder setStatus(boolean value) {
status_ = value;
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>optional bool status = 1;</code>
* @return This builder for chaining.
*/
public Builder clearStatus() {
bitField0_ = (bitField0_ & ~0x00000001);
status_ = false;
onChanged();
return this;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.RefreshSuperUserGroupsConfigurationResponseProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.RefreshSuperUserGroupsConfigurationResponseProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationResponseProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationResponseProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationResponseProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<RefreshSuperUserGroupsConfigurationResponseProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<RefreshSuperUserGroupsConfigurationResponseProto>() {
@java.lang.Override
public RefreshSuperUserGroupsConfigurationResponseProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<RefreshSuperUserGroupsConfigurationResponseProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<RefreshSuperUserGroupsConfigurationResponseProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.RefreshSuperUserGroupsConfigurationResponseProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface EnterSafeModeRequestProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.EnterSafeModeRequestProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
}
/**
* Protobuf type {@code hadoop.hdfs.EnterSafeModeRequestProto}
*/
public static final class EnterSafeModeRequestProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.EnterSafeModeRequestProto)
EnterSafeModeRequestProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use EnterSafeModeRequestProto.newBuilder() to construct.
private EnterSafeModeRequestProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private EnterSafeModeRequestProto() {
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new EnterSafeModeRequestProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_EnterSafeModeRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_EnterSafeModeRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeRequestProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeRequestProto.Builder.class);
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeRequestProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeRequestProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeRequestProto) obj;
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeRequestProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeRequestProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeRequestProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeRequestProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeRequestProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeRequestProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeRequestProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeRequestProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeRequestProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.EnterSafeModeRequestProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.EnterSafeModeRequestProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeRequestProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_EnterSafeModeRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_EnterSafeModeRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeRequestProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeRequestProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeRequestProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_EnterSafeModeRequestProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeRequestProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeRequestProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeRequestProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeRequestProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeRequestProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeRequestProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeRequestProto(this);
onBuilt();
return result;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeRequestProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeRequestProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeRequestProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeRequestProto.getDefaultInstance()) return this;
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.EnterSafeModeRequestProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.EnterSafeModeRequestProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeRequestProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeRequestProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeRequestProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<EnterSafeModeRequestProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<EnterSafeModeRequestProto>() {
@java.lang.Override
public EnterSafeModeRequestProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<EnterSafeModeRequestProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<EnterSafeModeRequestProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeRequestProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface EnterSafeModeResponseProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.EnterSafeModeResponseProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* <code>optional bool status = 1;</code>
* @return Whether the status field is set.
*/
boolean hasStatus();
/**
* <code>optional bool status = 1;</code>
* @return The status.
*/
boolean getStatus();
}
/**
* Protobuf type {@code hadoop.hdfs.EnterSafeModeResponseProto}
*/
public static final class EnterSafeModeResponseProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.EnterSafeModeResponseProto)
EnterSafeModeResponseProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use EnterSafeModeResponseProto.newBuilder() to construct.
private EnterSafeModeResponseProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private EnterSafeModeResponseProto() {
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new EnterSafeModeResponseProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_EnterSafeModeResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_EnterSafeModeResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeResponseProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeResponseProto.Builder.class);
}
private int bitField0_;
public static final int STATUS_FIELD_NUMBER = 1;
private boolean status_ = false;
/**
* <code>optional bool status = 1;</code>
* @return Whether the status field is set.
*/
@java.lang.Override
public boolean hasStatus() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional bool status = 1;</code>
* @return The status.
*/
@java.lang.Override
public boolean getStatus() {
return status_;
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) != 0)) {
output.writeBool(1, status_);
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeBoolSize(1, status_);
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeResponseProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeResponseProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeResponseProto) obj;
if (hasStatus() != other.hasStatus()) return false;
if (hasStatus()) {
if (getStatus()
!= other.getStatus()) return false;
}
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (hasStatus()) {
hash = (37 * hash) + STATUS_FIELD_NUMBER;
hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashBoolean(
getStatus());
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeResponseProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeResponseProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeResponseProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeResponseProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeResponseProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeResponseProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeResponseProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeResponseProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeResponseProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.EnterSafeModeResponseProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.EnterSafeModeResponseProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeResponseProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_EnterSafeModeResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_EnterSafeModeResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeResponseProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeResponseProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeResponseProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
status_ = false;
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_EnterSafeModeResponseProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeResponseProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeResponseProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeResponseProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeResponseProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeResponseProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeResponseProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeResponseProto(this);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartial0(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeResponseProto result) {
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) != 0)) {
result.status_ = status_;
to_bitField0_ |= 0x00000001;
}
result.bitField0_ |= to_bitField0_;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeResponseProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeResponseProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeResponseProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeResponseProto.getDefaultInstance()) return this;
if (other.hasStatus()) {
setStatus(other.getStatus());
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 8: {
status_ = input.readBool();
bitField0_ |= 0x00000001;
break;
} // case 8
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private boolean status_ ;
/**
* <code>optional bool status = 1;</code>
* @return Whether the status field is set.
*/
@java.lang.Override
public boolean hasStatus() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional bool status = 1;</code>
* @return The status.
*/
@java.lang.Override
public boolean getStatus() {
return status_;
}
/**
* <code>optional bool status = 1;</code>
* @param value The status to set.
* @return This builder for chaining.
*/
public Builder setStatus(boolean value) {
status_ = value;
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>optional bool status = 1;</code>
* @return This builder for chaining.
*/
public Builder clearStatus() {
bitField0_ = (bitField0_ & ~0x00000001);
status_ = false;
onChanged();
return this;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.EnterSafeModeResponseProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.EnterSafeModeResponseProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeResponseProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeResponseProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeResponseProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<EnterSafeModeResponseProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<EnterSafeModeResponseProto>() {
@java.lang.Override
public EnterSafeModeResponseProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<EnterSafeModeResponseProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<EnterSafeModeResponseProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnterSafeModeResponseProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface LeaveSafeModeRequestProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.LeaveSafeModeRequestProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
}
/**
* Protobuf type {@code hadoop.hdfs.LeaveSafeModeRequestProto}
*/
public static final class LeaveSafeModeRequestProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.LeaveSafeModeRequestProto)
LeaveSafeModeRequestProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use LeaveSafeModeRequestProto.newBuilder() to construct.
private LeaveSafeModeRequestProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private LeaveSafeModeRequestProto() {
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new LeaveSafeModeRequestProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_LeaveSafeModeRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_LeaveSafeModeRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeRequestProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeRequestProto.Builder.class);
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeRequestProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeRequestProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeRequestProto) obj;
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeRequestProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeRequestProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeRequestProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeRequestProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeRequestProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeRequestProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeRequestProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeRequestProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeRequestProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.LeaveSafeModeRequestProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.LeaveSafeModeRequestProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeRequestProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_LeaveSafeModeRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_LeaveSafeModeRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeRequestProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeRequestProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeRequestProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_LeaveSafeModeRequestProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeRequestProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeRequestProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeRequestProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeRequestProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeRequestProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeRequestProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeRequestProto(this);
onBuilt();
return result;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeRequestProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeRequestProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeRequestProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeRequestProto.getDefaultInstance()) return this;
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.LeaveSafeModeRequestProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.LeaveSafeModeRequestProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeRequestProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeRequestProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeRequestProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<LeaveSafeModeRequestProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<LeaveSafeModeRequestProto>() {
@java.lang.Override
public LeaveSafeModeRequestProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<LeaveSafeModeRequestProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<LeaveSafeModeRequestProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeRequestProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface LeaveSafeModeResponseProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.LeaveSafeModeResponseProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* <code>optional bool status = 1;</code>
* @return Whether the status field is set.
*/
boolean hasStatus();
/**
* <code>optional bool status = 1;</code>
* @return The status.
*/
boolean getStatus();
}
/**
* Protobuf type {@code hadoop.hdfs.LeaveSafeModeResponseProto}
*/
public static final class LeaveSafeModeResponseProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.LeaveSafeModeResponseProto)
LeaveSafeModeResponseProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use LeaveSafeModeResponseProto.newBuilder() to construct.
private LeaveSafeModeResponseProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private LeaveSafeModeResponseProto() {
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new LeaveSafeModeResponseProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_LeaveSafeModeResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_LeaveSafeModeResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeResponseProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeResponseProto.Builder.class);
}
private int bitField0_;
public static final int STATUS_FIELD_NUMBER = 1;
private boolean status_ = false;
/**
* <code>optional bool status = 1;</code>
* @return Whether the status field is set.
*/
@java.lang.Override
public boolean hasStatus() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional bool status = 1;</code>
* @return The status.
*/
@java.lang.Override
public boolean getStatus() {
return status_;
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) != 0)) {
output.writeBool(1, status_);
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeBoolSize(1, status_);
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeResponseProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeResponseProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeResponseProto) obj;
if (hasStatus() != other.hasStatus()) return false;
if (hasStatus()) {
if (getStatus()
!= other.getStatus()) return false;
}
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (hasStatus()) {
hash = (37 * hash) + STATUS_FIELD_NUMBER;
hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashBoolean(
getStatus());
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeResponseProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeResponseProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeResponseProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeResponseProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeResponseProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeResponseProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeResponseProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeResponseProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeResponseProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.LeaveSafeModeResponseProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.LeaveSafeModeResponseProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeResponseProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_LeaveSafeModeResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_LeaveSafeModeResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeResponseProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeResponseProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeResponseProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
status_ = false;
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_LeaveSafeModeResponseProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeResponseProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeResponseProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeResponseProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeResponseProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeResponseProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeResponseProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeResponseProto(this);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartial0(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeResponseProto result) {
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) != 0)) {
result.status_ = status_;
to_bitField0_ |= 0x00000001;
}
result.bitField0_ |= to_bitField0_;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeResponseProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeResponseProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeResponseProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeResponseProto.getDefaultInstance()) return this;
if (other.hasStatus()) {
setStatus(other.getStatus());
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 8: {
status_ = input.readBool();
bitField0_ |= 0x00000001;
break;
} // case 8
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private boolean status_ ;
/**
* <code>optional bool status = 1;</code>
* @return Whether the status field is set.
*/
@java.lang.Override
public boolean hasStatus() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional bool status = 1;</code>
* @return The status.
*/
@java.lang.Override
public boolean getStatus() {
return status_;
}
/**
* <code>optional bool status = 1;</code>
* @param value The status to set.
* @return This builder for chaining.
*/
public Builder setStatus(boolean value) {
status_ = value;
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>optional bool status = 1;</code>
* @return This builder for chaining.
*/
public Builder clearStatus() {
bitField0_ = (bitField0_ & ~0x00000001);
status_ = false;
onChanged();
return this;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.LeaveSafeModeResponseProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.LeaveSafeModeResponseProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeResponseProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeResponseProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeResponseProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<LeaveSafeModeResponseProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<LeaveSafeModeResponseProto>() {
@java.lang.Override
public LeaveSafeModeResponseProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<LeaveSafeModeResponseProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<LeaveSafeModeResponseProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.LeaveSafeModeResponseProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface GetSafeModeRequestProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.GetSafeModeRequestProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
}
/**
* Protobuf type {@code hadoop.hdfs.GetSafeModeRequestProto}
*/
public static final class GetSafeModeRequestProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.GetSafeModeRequestProto)
GetSafeModeRequestProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use GetSafeModeRequestProto.newBuilder() to construct.
private GetSafeModeRequestProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private GetSafeModeRequestProto() {
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new GetSafeModeRequestProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetSafeModeRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetSafeModeRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeRequestProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeRequestProto.Builder.class);
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeRequestProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeRequestProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeRequestProto) obj;
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeRequestProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeRequestProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeRequestProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeRequestProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeRequestProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeRequestProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeRequestProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeRequestProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeRequestProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.GetSafeModeRequestProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.GetSafeModeRequestProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeRequestProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetSafeModeRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetSafeModeRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeRequestProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeRequestProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeRequestProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetSafeModeRequestProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeRequestProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeRequestProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeRequestProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeRequestProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeRequestProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeRequestProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeRequestProto(this);
onBuilt();
return result;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeRequestProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeRequestProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeRequestProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeRequestProto.getDefaultInstance()) return this;
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.GetSafeModeRequestProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.GetSafeModeRequestProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeRequestProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeRequestProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeRequestProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<GetSafeModeRequestProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<GetSafeModeRequestProto>() {
@java.lang.Override
public GetSafeModeRequestProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<GetSafeModeRequestProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<GetSafeModeRequestProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeRequestProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface GetSafeModeResponseProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.GetSafeModeResponseProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* <code>optional bool isInSafeMode = 1;</code>
* @return Whether the isInSafeMode field is set.
*/
boolean hasIsInSafeMode();
/**
* <code>optional bool isInSafeMode = 1;</code>
* @return The isInSafeMode.
*/
boolean getIsInSafeMode();
}
/**
* Protobuf type {@code hadoop.hdfs.GetSafeModeResponseProto}
*/
public static final class GetSafeModeResponseProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.GetSafeModeResponseProto)
GetSafeModeResponseProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use GetSafeModeResponseProto.newBuilder() to construct.
private GetSafeModeResponseProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private GetSafeModeResponseProto() {
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new GetSafeModeResponseProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetSafeModeResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetSafeModeResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeResponseProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeResponseProto.Builder.class);
}
private int bitField0_;
public static final int ISINSAFEMODE_FIELD_NUMBER = 1;
private boolean isInSafeMode_ = false;
/**
* <code>optional bool isInSafeMode = 1;</code>
* @return Whether the isInSafeMode field is set.
*/
@java.lang.Override
public boolean hasIsInSafeMode() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional bool isInSafeMode = 1;</code>
* @return The isInSafeMode.
*/
@java.lang.Override
public boolean getIsInSafeMode() {
return isInSafeMode_;
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) != 0)) {
output.writeBool(1, isInSafeMode_);
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeBoolSize(1, isInSafeMode_);
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeResponseProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeResponseProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeResponseProto) obj;
if (hasIsInSafeMode() != other.hasIsInSafeMode()) return false;
if (hasIsInSafeMode()) {
if (getIsInSafeMode()
!= other.getIsInSafeMode()) return false;
}
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (hasIsInSafeMode()) {
hash = (37 * hash) + ISINSAFEMODE_FIELD_NUMBER;
hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashBoolean(
getIsInSafeMode());
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeResponseProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeResponseProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeResponseProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeResponseProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeResponseProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeResponseProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeResponseProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeResponseProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeResponseProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.GetSafeModeResponseProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.GetSafeModeResponseProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeResponseProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetSafeModeResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetSafeModeResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeResponseProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeResponseProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeResponseProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
isInSafeMode_ = false;
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetSafeModeResponseProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeResponseProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeResponseProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeResponseProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeResponseProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeResponseProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeResponseProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeResponseProto(this);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartial0(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeResponseProto result) {
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) != 0)) {
result.isInSafeMode_ = isInSafeMode_;
to_bitField0_ |= 0x00000001;
}
result.bitField0_ |= to_bitField0_;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeResponseProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeResponseProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeResponseProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeResponseProto.getDefaultInstance()) return this;
if (other.hasIsInSafeMode()) {
setIsInSafeMode(other.getIsInSafeMode());
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 8: {
isInSafeMode_ = input.readBool();
bitField0_ |= 0x00000001;
break;
} // case 8
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private boolean isInSafeMode_ ;
/**
* <code>optional bool isInSafeMode = 1;</code>
* @return Whether the isInSafeMode field is set.
*/
@java.lang.Override
public boolean hasIsInSafeMode() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional bool isInSafeMode = 1;</code>
* @return The isInSafeMode.
*/
@java.lang.Override
public boolean getIsInSafeMode() {
return isInSafeMode_;
}
/**
* <code>optional bool isInSafeMode = 1;</code>
* @param value The isInSafeMode to set.
* @return This builder for chaining.
*/
public Builder setIsInSafeMode(boolean value) {
isInSafeMode_ = value;
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>optional bool isInSafeMode = 1;</code>
* @return This builder for chaining.
*/
public Builder clearIsInSafeMode() {
bitField0_ = (bitField0_ & ~0x00000001);
isInSafeMode_ = false;
onChanged();
return this;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.GetSafeModeResponseProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.GetSafeModeResponseProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeResponseProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeResponseProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeResponseProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<GetSafeModeResponseProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<GetSafeModeResponseProto>() {
@java.lang.Override
public GetSafeModeResponseProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<GetSafeModeResponseProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<GetSafeModeResponseProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetSafeModeResponseProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface DisabledNameserviceRecordProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.DisabledNameserviceRecordProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* <code>optional uint64 dateCreated = 1;</code>
* @return Whether the dateCreated field is set.
*/
boolean hasDateCreated();
/**
* <code>optional uint64 dateCreated = 1;</code>
* @return The dateCreated.
*/
long getDateCreated();
/**
* <code>optional uint64 dateModified = 2;</code>
* @return Whether the dateModified field is set.
*/
boolean hasDateModified();
/**
* <code>optional uint64 dateModified = 2;</code>
* @return The dateModified.
*/
long getDateModified();
/**
* <code>optional string nameServiceId = 3;</code>
* @return Whether the nameServiceId field is set.
*/
boolean hasNameServiceId();
/**
* <code>optional string nameServiceId = 3;</code>
* @return The nameServiceId.
*/
java.lang.String getNameServiceId();
/**
* <code>optional string nameServiceId = 3;</code>
* @return The bytes for nameServiceId.
*/
org.apache.hadoop.thirdparty.protobuf.ByteString
getNameServiceIdBytes();
}
/**
* Protobuf type {@code hadoop.hdfs.DisabledNameserviceRecordProto}
*/
public static final class DisabledNameserviceRecordProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.DisabledNameserviceRecordProto)
DisabledNameserviceRecordProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use DisabledNameserviceRecordProto.newBuilder() to construct.
private DisabledNameserviceRecordProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private DisabledNameserviceRecordProto() {
nameServiceId_ = "";
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new DisabledNameserviceRecordProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_DisabledNameserviceRecordProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_DisabledNameserviceRecordProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisabledNameserviceRecordProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisabledNameserviceRecordProto.Builder.class);
}
private int bitField0_;
public static final int DATECREATED_FIELD_NUMBER = 1;
private long dateCreated_ = 0L;
/**
* <code>optional uint64 dateCreated = 1;</code>
* @return Whether the dateCreated field is set.
*/
@java.lang.Override
public boolean hasDateCreated() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional uint64 dateCreated = 1;</code>
* @return The dateCreated.
*/
@java.lang.Override
public long getDateCreated() {
return dateCreated_;
}
public static final int DATEMODIFIED_FIELD_NUMBER = 2;
private long dateModified_ = 0L;
/**
* <code>optional uint64 dateModified = 2;</code>
* @return Whether the dateModified field is set.
*/
@java.lang.Override
public boolean hasDateModified() {
return ((bitField0_ & 0x00000002) != 0);
}
/**
* <code>optional uint64 dateModified = 2;</code>
* @return The dateModified.
*/
@java.lang.Override
public long getDateModified() {
return dateModified_;
}
public static final int NAMESERVICEID_FIELD_NUMBER = 3;
@SuppressWarnings("serial")
private volatile java.lang.Object nameServiceId_ = "";
/**
* <code>optional string nameServiceId = 3;</code>
* @return Whether the nameServiceId field is set.
*/
@java.lang.Override
public boolean hasNameServiceId() {
return ((bitField0_ & 0x00000004) != 0);
}
/**
* <code>optional string nameServiceId = 3;</code>
* @return The nameServiceId.
*/
@java.lang.Override
public java.lang.String getNameServiceId() {
java.lang.Object ref = nameServiceId_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
nameServiceId_ = s;
}
return s;
}
}
/**
* <code>optional string nameServiceId = 3;</code>
* @return The bytes for nameServiceId.
*/
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.ByteString
getNameServiceIdBytes() {
java.lang.Object ref = nameServiceId_;
if (ref instanceof java.lang.String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
nameServiceId_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) != 0)) {
output.writeUInt64(1, dateCreated_);
}
if (((bitField0_ & 0x00000002) != 0)) {
output.writeUInt64(2, dateModified_);
}
if (((bitField0_ & 0x00000004) != 0)) {
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 3, nameServiceId_);
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeUInt64Size(1, dateCreated_);
}
if (((bitField0_ & 0x00000002) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeUInt64Size(2, dateModified_);
}
if (((bitField0_ & 0x00000004) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(3, nameServiceId_);
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisabledNameserviceRecordProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisabledNameserviceRecordProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisabledNameserviceRecordProto) obj;
if (hasDateCreated() != other.hasDateCreated()) return false;
if (hasDateCreated()) {
if (getDateCreated()
!= other.getDateCreated()) return false;
}
if (hasDateModified() != other.hasDateModified()) return false;
if (hasDateModified()) {
if (getDateModified()
!= other.getDateModified()) return false;
}
if (hasNameServiceId() != other.hasNameServiceId()) return false;
if (hasNameServiceId()) {
if (!getNameServiceId()
.equals(other.getNameServiceId())) return false;
}
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (hasDateCreated()) {
hash = (37 * hash) + DATECREATED_FIELD_NUMBER;
hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashLong(
getDateCreated());
}
if (hasDateModified()) {
hash = (37 * hash) + DATEMODIFIED_FIELD_NUMBER;
hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashLong(
getDateModified());
}
if (hasNameServiceId()) {
hash = (37 * hash) + NAMESERVICEID_FIELD_NUMBER;
hash = (53 * hash) + getNameServiceId().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisabledNameserviceRecordProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisabledNameserviceRecordProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisabledNameserviceRecordProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisabledNameserviceRecordProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisabledNameserviceRecordProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisabledNameserviceRecordProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisabledNameserviceRecordProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisabledNameserviceRecordProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisabledNameserviceRecordProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisabledNameserviceRecordProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisabledNameserviceRecordProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisabledNameserviceRecordProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisabledNameserviceRecordProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.DisabledNameserviceRecordProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.DisabledNameserviceRecordProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisabledNameserviceRecordProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_DisabledNameserviceRecordProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_DisabledNameserviceRecordProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisabledNameserviceRecordProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisabledNameserviceRecordProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisabledNameserviceRecordProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
dateCreated_ = 0L;
dateModified_ = 0L;
nameServiceId_ = "";
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_DisabledNameserviceRecordProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisabledNameserviceRecordProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisabledNameserviceRecordProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisabledNameserviceRecordProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisabledNameserviceRecordProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisabledNameserviceRecordProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisabledNameserviceRecordProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisabledNameserviceRecordProto(this);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartial0(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisabledNameserviceRecordProto result) {
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) != 0)) {
result.dateCreated_ = dateCreated_;
to_bitField0_ |= 0x00000001;
}
if (((from_bitField0_ & 0x00000002) != 0)) {
result.dateModified_ = dateModified_;
to_bitField0_ |= 0x00000002;
}
if (((from_bitField0_ & 0x00000004) != 0)) {
result.nameServiceId_ = nameServiceId_;
to_bitField0_ |= 0x00000004;
}
result.bitField0_ |= to_bitField0_;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisabledNameserviceRecordProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisabledNameserviceRecordProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisabledNameserviceRecordProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisabledNameserviceRecordProto.getDefaultInstance()) return this;
if (other.hasDateCreated()) {
setDateCreated(other.getDateCreated());
}
if (other.hasDateModified()) {
setDateModified(other.getDateModified());
}
if (other.hasNameServiceId()) {
nameServiceId_ = other.nameServiceId_;
bitField0_ |= 0x00000004;
onChanged();
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 8: {
dateCreated_ = input.readUInt64();
bitField0_ |= 0x00000001;
break;
} // case 8
case 16: {
dateModified_ = input.readUInt64();
bitField0_ |= 0x00000002;
break;
} // case 16
case 26: {
nameServiceId_ = input.readBytes();
bitField0_ |= 0x00000004;
break;
} // case 26
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private long dateCreated_ ;
/**
* <code>optional uint64 dateCreated = 1;</code>
* @return Whether the dateCreated field is set.
*/
@java.lang.Override
public boolean hasDateCreated() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional uint64 dateCreated = 1;</code>
* @return The dateCreated.
*/
@java.lang.Override
public long getDateCreated() {
return dateCreated_;
}
/**
* <code>optional uint64 dateCreated = 1;</code>
* @param value The dateCreated to set.
* @return This builder for chaining.
*/
public Builder setDateCreated(long value) {
dateCreated_ = value;
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>optional uint64 dateCreated = 1;</code>
* @return This builder for chaining.
*/
public Builder clearDateCreated() {
bitField0_ = (bitField0_ & ~0x00000001);
dateCreated_ = 0L;
onChanged();
return this;
}
private long dateModified_ ;
/**
* <code>optional uint64 dateModified = 2;</code>
* @return Whether the dateModified field is set.
*/
@java.lang.Override
public boolean hasDateModified() {
return ((bitField0_ & 0x00000002) != 0);
}
/**
* <code>optional uint64 dateModified = 2;</code>
* @return The dateModified.
*/
@java.lang.Override
public long getDateModified() {
return dateModified_;
}
/**
* <code>optional uint64 dateModified = 2;</code>
* @param value The dateModified to set.
* @return This builder for chaining.
*/
public Builder setDateModified(long value) {
dateModified_ = value;
bitField0_ |= 0x00000002;
onChanged();
return this;
}
/**
* <code>optional uint64 dateModified = 2;</code>
* @return This builder for chaining.
*/
public Builder clearDateModified() {
bitField0_ = (bitField0_ & ~0x00000002);
dateModified_ = 0L;
onChanged();
return this;
}
private java.lang.Object nameServiceId_ = "";
/**
* <code>optional string nameServiceId = 3;</code>
* @return Whether the nameServiceId field is set.
*/
public boolean hasNameServiceId() {
return ((bitField0_ & 0x00000004) != 0);
}
/**
* <code>optional string nameServiceId = 3;</code>
* @return The nameServiceId.
*/
public java.lang.String getNameServiceId() {
java.lang.Object ref = nameServiceId_;
if (!(ref instanceof java.lang.String)) {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
nameServiceId_ = s;
}
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* <code>optional string nameServiceId = 3;</code>
* @return The bytes for nameServiceId.
*/
public org.apache.hadoop.thirdparty.protobuf.ByteString
getNameServiceIdBytes() {
java.lang.Object ref = nameServiceId_;
if (ref instanceof String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
nameServiceId_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
/**
* <code>optional string nameServiceId = 3;</code>
* @param value The nameServiceId to set.
* @return This builder for chaining.
*/
public Builder setNameServiceId(
java.lang.String value) {
if (value == null) { throw new NullPointerException(); }
nameServiceId_ = value;
bitField0_ |= 0x00000004;
onChanged();
return this;
}
/**
* <code>optional string nameServiceId = 3;</code>
* @return This builder for chaining.
*/
public Builder clearNameServiceId() {
nameServiceId_ = getDefaultInstance().getNameServiceId();
bitField0_ = (bitField0_ & ~0x00000004);
onChanged();
return this;
}
/**
* <code>optional string nameServiceId = 3;</code>
* @param value The bytes for nameServiceId to set.
* @return This builder for chaining.
*/
public Builder setNameServiceIdBytes(
org.apache.hadoop.thirdparty.protobuf.ByteString value) {
if (value == null) { throw new NullPointerException(); }
nameServiceId_ = value;
bitField0_ |= 0x00000004;
onChanged();
return this;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.DisabledNameserviceRecordProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.DisabledNameserviceRecordProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisabledNameserviceRecordProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisabledNameserviceRecordProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisabledNameserviceRecordProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<DisabledNameserviceRecordProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<DisabledNameserviceRecordProto>() {
@java.lang.Override
public DisabledNameserviceRecordProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<DisabledNameserviceRecordProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<DisabledNameserviceRecordProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisabledNameserviceRecordProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface DisableNameserviceRequestProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.DisableNameserviceRequestProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* <code>optional string nameServiceId = 1;</code>
* @return Whether the nameServiceId field is set.
*/
boolean hasNameServiceId();
/**
* <code>optional string nameServiceId = 1;</code>
* @return The nameServiceId.
*/
java.lang.String getNameServiceId();
/**
* <code>optional string nameServiceId = 1;</code>
* @return The bytes for nameServiceId.
*/
org.apache.hadoop.thirdparty.protobuf.ByteString
getNameServiceIdBytes();
}
/**
* Protobuf type {@code hadoop.hdfs.DisableNameserviceRequestProto}
*/
public static final class DisableNameserviceRequestProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.DisableNameserviceRequestProto)
DisableNameserviceRequestProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use DisableNameserviceRequestProto.newBuilder() to construct.
private DisableNameserviceRequestProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private DisableNameserviceRequestProto() {
nameServiceId_ = "";
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new DisableNameserviceRequestProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_DisableNameserviceRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_DisableNameserviceRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceRequestProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceRequestProto.Builder.class);
}
private int bitField0_;
public static final int NAMESERVICEID_FIELD_NUMBER = 1;
@SuppressWarnings("serial")
private volatile java.lang.Object nameServiceId_ = "";
/**
* <code>optional string nameServiceId = 1;</code>
* @return Whether the nameServiceId field is set.
*/
@java.lang.Override
public boolean hasNameServiceId() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional string nameServiceId = 1;</code>
* @return The nameServiceId.
*/
@java.lang.Override
public java.lang.String getNameServiceId() {
java.lang.Object ref = nameServiceId_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
nameServiceId_ = s;
}
return s;
}
}
/**
* <code>optional string nameServiceId = 1;</code>
* @return The bytes for nameServiceId.
*/
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.ByteString
getNameServiceIdBytes() {
java.lang.Object ref = nameServiceId_;
if (ref instanceof java.lang.String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
nameServiceId_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) != 0)) {
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 1, nameServiceId_);
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(1, nameServiceId_);
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceRequestProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceRequestProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceRequestProto) obj;
if (hasNameServiceId() != other.hasNameServiceId()) return false;
if (hasNameServiceId()) {
if (!getNameServiceId()
.equals(other.getNameServiceId())) return false;
}
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (hasNameServiceId()) {
hash = (37 * hash) + NAMESERVICEID_FIELD_NUMBER;
hash = (53 * hash) + getNameServiceId().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceRequestProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceRequestProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceRequestProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceRequestProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceRequestProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceRequestProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceRequestProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceRequestProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceRequestProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.DisableNameserviceRequestProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.DisableNameserviceRequestProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceRequestProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_DisableNameserviceRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_DisableNameserviceRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceRequestProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceRequestProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceRequestProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
nameServiceId_ = "";
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_DisableNameserviceRequestProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceRequestProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceRequestProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceRequestProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceRequestProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceRequestProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceRequestProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceRequestProto(this);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartial0(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceRequestProto result) {
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) != 0)) {
result.nameServiceId_ = nameServiceId_;
to_bitField0_ |= 0x00000001;
}
result.bitField0_ |= to_bitField0_;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceRequestProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceRequestProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceRequestProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceRequestProto.getDefaultInstance()) return this;
if (other.hasNameServiceId()) {
nameServiceId_ = other.nameServiceId_;
bitField0_ |= 0x00000001;
onChanged();
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 10: {
nameServiceId_ = input.readBytes();
bitField0_ |= 0x00000001;
break;
} // case 10
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private java.lang.Object nameServiceId_ = "";
/**
* <code>optional string nameServiceId = 1;</code>
* @return Whether the nameServiceId field is set.
*/
public boolean hasNameServiceId() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional string nameServiceId = 1;</code>
* @return The nameServiceId.
*/
public java.lang.String getNameServiceId() {
java.lang.Object ref = nameServiceId_;
if (!(ref instanceof java.lang.String)) {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
nameServiceId_ = s;
}
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* <code>optional string nameServiceId = 1;</code>
* @return The bytes for nameServiceId.
*/
public org.apache.hadoop.thirdparty.protobuf.ByteString
getNameServiceIdBytes() {
java.lang.Object ref = nameServiceId_;
if (ref instanceof String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
nameServiceId_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
/**
* <code>optional string nameServiceId = 1;</code>
* @param value The nameServiceId to set.
* @return This builder for chaining.
*/
public Builder setNameServiceId(
java.lang.String value) {
if (value == null) { throw new NullPointerException(); }
nameServiceId_ = value;
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>optional string nameServiceId = 1;</code>
* @return This builder for chaining.
*/
public Builder clearNameServiceId() {
nameServiceId_ = getDefaultInstance().getNameServiceId();
bitField0_ = (bitField0_ & ~0x00000001);
onChanged();
return this;
}
/**
* <code>optional string nameServiceId = 1;</code>
* @param value The bytes for nameServiceId to set.
* @return This builder for chaining.
*/
public Builder setNameServiceIdBytes(
org.apache.hadoop.thirdparty.protobuf.ByteString value) {
if (value == null) { throw new NullPointerException(); }
nameServiceId_ = value;
bitField0_ |= 0x00000001;
onChanged();
return this;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.DisableNameserviceRequestProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.DisableNameserviceRequestProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceRequestProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceRequestProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceRequestProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<DisableNameserviceRequestProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<DisableNameserviceRequestProto>() {
@java.lang.Override
public DisableNameserviceRequestProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<DisableNameserviceRequestProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<DisableNameserviceRequestProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceRequestProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface DisableNameserviceResponseProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.DisableNameserviceResponseProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* <code>optional bool status = 1;</code>
* @return Whether the status field is set.
*/
boolean hasStatus();
/**
* <code>optional bool status = 1;</code>
* @return The status.
*/
boolean getStatus();
}
/**
* Protobuf type {@code hadoop.hdfs.DisableNameserviceResponseProto}
*/
public static final class DisableNameserviceResponseProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.DisableNameserviceResponseProto)
DisableNameserviceResponseProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use DisableNameserviceResponseProto.newBuilder() to construct.
private DisableNameserviceResponseProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private DisableNameserviceResponseProto() {
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new DisableNameserviceResponseProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_DisableNameserviceResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_DisableNameserviceResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceResponseProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceResponseProto.Builder.class);
}
private int bitField0_;
public static final int STATUS_FIELD_NUMBER = 1;
private boolean status_ = false;
/**
* <code>optional bool status = 1;</code>
* @return Whether the status field is set.
*/
@java.lang.Override
public boolean hasStatus() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional bool status = 1;</code>
* @return The status.
*/
@java.lang.Override
public boolean getStatus() {
return status_;
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) != 0)) {
output.writeBool(1, status_);
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeBoolSize(1, status_);
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceResponseProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceResponseProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceResponseProto) obj;
if (hasStatus() != other.hasStatus()) return false;
if (hasStatus()) {
if (getStatus()
!= other.getStatus()) return false;
}
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (hasStatus()) {
hash = (37 * hash) + STATUS_FIELD_NUMBER;
hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashBoolean(
getStatus());
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceResponseProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceResponseProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceResponseProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceResponseProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceResponseProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceResponseProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceResponseProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceResponseProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceResponseProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.DisableNameserviceResponseProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.DisableNameserviceResponseProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceResponseProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_DisableNameserviceResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_DisableNameserviceResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceResponseProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceResponseProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceResponseProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
status_ = false;
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_DisableNameserviceResponseProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceResponseProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceResponseProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceResponseProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceResponseProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceResponseProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceResponseProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceResponseProto(this);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartial0(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceResponseProto result) {
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) != 0)) {
result.status_ = status_;
to_bitField0_ |= 0x00000001;
}
result.bitField0_ |= to_bitField0_;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceResponseProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceResponseProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceResponseProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceResponseProto.getDefaultInstance()) return this;
if (other.hasStatus()) {
setStatus(other.getStatus());
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 8: {
status_ = input.readBool();
bitField0_ |= 0x00000001;
break;
} // case 8
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private boolean status_ ;
/**
* <code>optional bool status = 1;</code>
* @return Whether the status field is set.
*/
@java.lang.Override
public boolean hasStatus() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional bool status = 1;</code>
* @return The status.
*/
@java.lang.Override
public boolean getStatus() {
return status_;
}
/**
* <code>optional bool status = 1;</code>
* @param value The status to set.
* @return This builder for chaining.
*/
public Builder setStatus(boolean value) {
status_ = value;
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>optional bool status = 1;</code>
* @return This builder for chaining.
*/
public Builder clearStatus() {
bitField0_ = (bitField0_ & ~0x00000001);
status_ = false;
onChanged();
return this;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.DisableNameserviceResponseProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.DisableNameserviceResponseProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceResponseProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceResponseProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceResponseProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<DisableNameserviceResponseProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<DisableNameserviceResponseProto>() {
@java.lang.Override
public DisableNameserviceResponseProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<DisableNameserviceResponseProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<DisableNameserviceResponseProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.DisableNameserviceResponseProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface EnableNameserviceRequestProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.EnableNameserviceRequestProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* <code>optional string nameServiceId = 1;</code>
* @return Whether the nameServiceId field is set.
*/
boolean hasNameServiceId();
/**
* <code>optional string nameServiceId = 1;</code>
* @return The nameServiceId.
*/
java.lang.String getNameServiceId();
/**
* <code>optional string nameServiceId = 1;</code>
* @return The bytes for nameServiceId.
*/
org.apache.hadoop.thirdparty.protobuf.ByteString
getNameServiceIdBytes();
}
/**
* Protobuf type {@code hadoop.hdfs.EnableNameserviceRequestProto}
*/
public static final class EnableNameserviceRequestProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.EnableNameserviceRequestProto)
EnableNameserviceRequestProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use EnableNameserviceRequestProto.newBuilder() to construct.
private EnableNameserviceRequestProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private EnableNameserviceRequestProto() {
nameServiceId_ = "";
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new EnableNameserviceRequestProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_EnableNameserviceRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_EnableNameserviceRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceRequestProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceRequestProto.Builder.class);
}
private int bitField0_;
public static final int NAMESERVICEID_FIELD_NUMBER = 1;
@SuppressWarnings("serial")
private volatile java.lang.Object nameServiceId_ = "";
/**
* <code>optional string nameServiceId = 1;</code>
* @return Whether the nameServiceId field is set.
*/
@java.lang.Override
public boolean hasNameServiceId() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional string nameServiceId = 1;</code>
* @return The nameServiceId.
*/
@java.lang.Override
public java.lang.String getNameServiceId() {
java.lang.Object ref = nameServiceId_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
nameServiceId_ = s;
}
return s;
}
}
/**
* <code>optional string nameServiceId = 1;</code>
* @return The bytes for nameServiceId.
*/
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.ByteString
getNameServiceIdBytes() {
java.lang.Object ref = nameServiceId_;
if (ref instanceof java.lang.String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
nameServiceId_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) != 0)) {
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 1, nameServiceId_);
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(1, nameServiceId_);
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceRequestProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceRequestProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceRequestProto) obj;
if (hasNameServiceId() != other.hasNameServiceId()) return false;
if (hasNameServiceId()) {
if (!getNameServiceId()
.equals(other.getNameServiceId())) return false;
}
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (hasNameServiceId()) {
hash = (37 * hash) + NAMESERVICEID_FIELD_NUMBER;
hash = (53 * hash) + getNameServiceId().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceRequestProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceRequestProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceRequestProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceRequestProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceRequestProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceRequestProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceRequestProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceRequestProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceRequestProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.EnableNameserviceRequestProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.EnableNameserviceRequestProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceRequestProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_EnableNameserviceRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_EnableNameserviceRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceRequestProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceRequestProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceRequestProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
nameServiceId_ = "";
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_EnableNameserviceRequestProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceRequestProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceRequestProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceRequestProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceRequestProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceRequestProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceRequestProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceRequestProto(this);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartial0(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceRequestProto result) {
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) != 0)) {
result.nameServiceId_ = nameServiceId_;
to_bitField0_ |= 0x00000001;
}
result.bitField0_ |= to_bitField0_;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceRequestProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceRequestProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceRequestProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceRequestProto.getDefaultInstance()) return this;
if (other.hasNameServiceId()) {
nameServiceId_ = other.nameServiceId_;
bitField0_ |= 0x00000001;
onChanged();
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 10: {
nameServiceId_ = input.readBytes();
bitField0_ |= 0x00000001;
break;
} // case 10
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private java.lang.Object nameServiceId_ = "";
/**
* <code>optional string nameServiceId = 1;</code>
* @return Whether the nameServiceId field is set.
*/
public boolean hasNameServiceId() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional string nameServiceId = 1;</code>
* @return The nameServiceId.
*/
public java.lang.String getNameServiceId() {
java.lang.Object ref = nameServiceId_;
if (!(ref instanceof java.lang.String)) {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
nameServiceId_ = s;
}
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* <code>optional string nameServiceId = 1;</code>
* @return The bytes for nameServiceId.
*/
public org.apache.hadoop.thirdparty.protobuf.ByteString
getNameServiceIdBytes() {
java.lang.Object ref = nameServiceId_;
if (ref instanceof String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
nameServiceId_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
/**
* <code>optional string nameServiceId = 1;</code>
* @param value The nameServiceId to set.
* @return This builder for chaining.
*/
public Builder setNameServiceId(
java.lang.String value) {
if (value == null) { throw new NullPointerException(); }
nameServiceId_ = value;
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>optional string nameServiceId = 1;</code>
* @return This builder for chaining.
*/
public Builder clearNameServiceId() {
nameServiceId_ = getDefaultInstance().getNameServiceId();
bitField0_ = (bitField0_ & ~0x00000001);
onChanged();
return this;
}
/**
* <code>optional string nameServiceId = 1;</code>
* @param value The bytes for nameServiceId to set.
* @return This builder for chaining.
*/
public Builder setNameServiceIdBytes(
org.apache.hadoop.thirdparty.protobuf.ByteString value) {
if (value == null) { throw new NullPointerException(); }
nameServiceId_ = value;
bitField0_ |= 0x00000001;
onChanged();
return this;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.EnableNameserviceRequestProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.EnableNameserviceRequestProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceRequestProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceRequestProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceRequestProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<EnableNameserviceRequestProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<EnableNameserviceRequestProto>() {
@java.lang.Override
public EnableNameserviceRequestProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<EnableNameserviceRequestProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<EnableNameserviceRequestProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceRequestProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface EnableNameserviceResponseProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.EnableNameserviceResponseProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* <code>optional bool status = 1;</code>
* @return Whether the status field is set.
*/
boolean hasStatus();
/**
* <code>optional bool status = 1;</code>
* @return The status.
*/
boolean getStatus();
}
/**
* Protobuf type {@code hadoop.hdfs.EnableNameserviceResponseProto}
*/
public static final class EnableNameserviceResponseProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.EnableNameserviceResponseProto)
EnableNameserviceResponseProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use EnableNameserviceResponseProto.newBuilder() to construct.
private EnableNameserviceResponseProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private EnableNameserviceResponseProto() {
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new EnableNameserviceResponseProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_EnableNameserviceResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_EnableNameserviceResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceResponseProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceResponseProto.Builder.class);
}
private int bitField0_;
public static final int STATUS_FIELD_NUMBER = 1;
private boolean status_ = false;
/**
* <code>optional bool status = 1;</code>
* @return Whether the status field is set.
*/
@java.lang.Override
public boolean hasStatus() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional bool status = 1;</code>
* @return The status.
*/
@java.lang.Override
public boolean getStatus() {
return status_;
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) != 0)) {
output.writeBool(1, status_);
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeBoolSize(1, status_);
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceResponseProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceResponseProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceResponseProto) obj;
if (hasStatus() != other.hasStatus()) return false;
if (hasStatus()) {
if (getStatus()
!= other.getStatus()) return false;
}
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (hasStatus()) {
hash = (37 * hash) + STATUS_FIELD_NUMBER;
hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashBoolean(
getStatus());
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceResponseProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceResponseProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceResponseProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceResponseProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceResponseProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceResponseProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceResponseProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceResponseProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceResponseProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.EnableNameserviceResponseProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.EnableNameserviceResponseProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceResponseProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_EnableNameserviceResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_EnableNameserviceResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceResponseProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceResponseProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceResponseProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
status_ = false;
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_EnableNameserviceResponseProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceResponseProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceResponseProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceResponseProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceResponseProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceResponseProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceResponseProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceResponseProto(this);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartial0(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceResponseProto result) {
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) != 0)) {
result.status_ = status_;
to_bitField0_ |= 0x00000001;
}
result.bitField0_ |= to_bitField0_;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceResponseProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceResponseProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceResponseProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceResponseProto.getDefaultInstance()) return this;
if (other.hasStatus()) {
setStatus(other.getStatus());
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 8: {
status_ = input.readBool();
bitField0_ |= 0x00000001;
break;
} // case 8
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private boolean status_ ;
/**
* <code>optional bool status = 1;</code>
* @return Whether the status field is set.
*/
@java.lang.Override
public boolean hasStatus() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* <code>optional bool status = 1;</code>
* @return The status.
*/
@java.lang.Override
public boolean getStatus() {
return status_;
}
/**
* <code>optional bool status = 1;</code>
* @param value The status to set.
* @return This builder for chaining.
*/
public Builder setStatus(boolean value) {
status_ = value;
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>optional bool status = 1;</code>
* @return This builder for chaining.
*/
public Builder clearStatus() {
bitField0_ = (bitField0_ & ~0x00000001);
status_ = false;
onChanged();
return this;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.EnableNameserviceResponseProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.EnableNameserviceResponseProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceResponseProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceResponseProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceResponseProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<EnableNameserviceResponseProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<EnableNameserviceResponseProto>() {
@java.lang.Override
public EnableNameserviceResponseProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<EnableNameserviceResponseProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<EnableNameserviceResponseProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.EnableNameserviceResponseProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface GetDisabledNameservicesRequestProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.GetDisabledNameservicesRequestProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
}
/**
* Protobuf type {@code hadoop.hdfs.GetDisabledNameservicesRequestProto}
*/
public static final class GetDisabledNameservicesRequestProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.GetDisabledNameservicesRequestProto)
GetDisabledNameservicesRequestProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use GetDisabledNameservicesRequestProto.newBuilder() to construct.
private GetDisabledNameservicesRequestProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private GetDisabledNameservicesRequestProto() {
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new GetDisabledNameservicesRequestProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetDisabledNameservicesRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetDisabledNameservicesRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesRequestProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesRequestProto.Builder.class);
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesRequestProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesRequestProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesRequestProto) obj;
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesRequestProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesRequestProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesRequestProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesRequestProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesRequestProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesRequestProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesRequestProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesRequestProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesRequestProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.GetDisabledNameservicesRequestProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.GetDisabledNameservicesRequestProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesRequestProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetDisabledNameservicesRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetDisabledNameservicesRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesRequestProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesRequestProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesRequestProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetDisabledNameservicesRequestProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesRequestProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesRequestProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesRequestProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesRequestProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesRequestProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesRequestProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesRequestProto(this);
onBuilt();
return result;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesRequestProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesRequestProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesRequestProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesRequestProto.getDefaultInstance()) return this;
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.GetDisabledNameservicesRequestProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.GetDisabledNameservicesRequestProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesRequestProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesRequestProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesRequestProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<GetDisabledNameservicesRequestProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<GetDisabledNameservicesRequestProto>() {
@java.lang.Override
public GetDisabledNameservicesRequestProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<GetDisabledNameservicesRequestProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<GetDisabledNameservicesRequestProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesRequestProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface GetDisabledNameservicesResponseProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.GetDisabledNameservicesResponseProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* <code>repeated string nameServiceIds = 1;</code>
* @return A list containing the nameServiceIds.
*/
java.util.List<java.lang.String>
getNameServiceIdsList();
/**
* <code>repeated string nameServiceIds = 1;</code>
* @return The count of nameServiceIds.
*/
int getNameServiceIdsCount();
/**
* <code>repeated string nameServiceIds = 1;</code>
* @param index The index of the element to return.
* @return The nameServiceIds at the given index.
*/
java.lang.String getNameServiceIds(int index);
/**
* <code>repeated string nameServiceIds = 1;</code>
* @param index The index of the value to return.
* @return The bytes of the nameServiceIds at the given index.
*/
org.apache.hadoop.thirdparty.protobuf.ByteString
getNameServiceIdsBytes(int index);
}
/**
* Protobuf type {@code hadoop.hdfs.GetDisabledNameservicesResponseProto}
*/
public static final class GetDisabledNameservicesResponseProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.GetDisabledNameservicesResponseProto)
GetDisabledNameservicesResponseProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use GetDisabledNameservicesResponseProto.newBuilder() to construct.
private GetDisabledNameservicesResponseProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<?> builder) {
super(builder);
}
private GetDisabledNameservicesResponseProto() {
nameServiceIds_ =
org.apache.hadoop.thirdparty.protobuf.LazyStringArrayList.emptyList();
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new GetDisabledNameservicesResponseProto();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetDisabledNameservicesResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetDisabledNameservicesResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesResponseProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesResponseProto.Builder.class);
}
public static final int NAMESERVICEIDS_FIELD_NUMBER = 1;
@SuppressWarnings("serial")
private org.apache.hadoop.thirdparty.protobuf.LazyStringArrayList nameServiceIds_ =
org.apache.hadoop.thirdparty.protobuf.LazyStringArrayList.emptyList();
/**
* <code>repeated string nameServiceIds = 1;</code>
* @return A list containing the nameServiceIds.
*/
public org.apache.hadoop.thirdparty.protobuf.ProtocolStringList
getNameServiceIdsList() {
return nameServiceIds_;
}
/**
* <code>repeated string nameServiceIds = 1;</code>
* @return The count of nameServiceIds.
*/
public int getNameServiceIdsCount() {
return nameServiceIds_.size();
}
/**
* <code>repeated string nameServiceIds = 1;</code>
* @param index The index of the element to return.
* @return The nameServiceIds at the given index.
*/
public java.lang.String getNameServiceIds(int index) {
return nameServiceIds_.get(index);
}
/**
* <code>repeated string nameServiceIds = 1;</code>
* @param index The index of the value to return.
* @return The bytes of the nameServiceIds at the given index.
*/
public org.apache.hadoop.thirdparty.protobuf.ByteString
getNameServiceIdsBytes(int index) {
return nameServiceIds_.getByteString(index);
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
for (int i = 0; i < nameServiceIds_.size(); i++) {
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 1, nameServiceIds_.getRaw(i));
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
{
int dataSize = 0;
for (int i = 0; i < nameServiceIds_.size(); i++) {
dataSize += computeStringSizeNoTag(nameServiceIds_.getRaw(i));
}
size += dataSize;
size += 1 * getNameServiceIdsList().size();
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesResponseProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesResponseProto other = (org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesResponseProto) obj;
if (!getNameServiceIdsList()
.equals(other.getNameServiceIdsList())) return false;
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (getNameServiceIdsCount() > 0) {
hash = (37 * hash) + NAMESERVICEIDS_FIELD_NUMBER;
hash = (53 * hash) + getNameServiceIdsList().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesResponseProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesResponseProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesResponseProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesResponseProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesResponseProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesResponseProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesResponseProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesResponseProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesResponseProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.GetDisabledNameservicesResponseProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder<Builder> implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.GetDisabledNameservicesResponseProto)
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesResponseProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetDisabledNameservicesResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetDisabledNameservicesResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesResponseProto.class, org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesResponseProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesResponseProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
nameServiceIds_ =
org.apache.hadoop.thirdparty.protobuf.LazyStringArrayList.emptyList();
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.internal_static_hadoop_hdfs_GetDisabledNameservicesResponseProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesResponseProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesResponseProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesResponseProto build() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesResponseProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesResponseProto buildPartial() {
org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesResponseProto result = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesResponseProto(this);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartial0(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesResponseProto result) {
int from_bitField0_ = bitField0_;
if (((from_bitField0_ & 0x00000001) != 0)) {
nameServiceIds_.makeImmutable();
result.nameServiceIds_ = nameServiceIds_;
}
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesResponseProto) {
return mergeFrom((org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesResponseProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesResponseProto other) {
if (other == org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesResponseProto.getDefaultInstance()) return this;
if (!other.nameServiceIds_.isEmpty()) {
if (nameServiceIds_.isEmpty()) {
nameServiceIds_ = other.nameServiceIds_;
bitField0_ |= 0x00000001;
} else {
ensureNameServiceIdsIsMutable();
nameServiceIds_.addAll(other.nameServiceIds_);
}
onChanged();
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 10: {
org.apache.hadoop.thirdparty.protobuf.ByteString bs = input.readBytes();
ensureNameServiceIdsIsMutable();
nameServiceIds_.add(bs);
break;
} // case 10
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private org.apache.hadoop.thirdparty.protobuf.LazyStringArrayList nameServiceIds_ =
org.apache.hadoop.thirdparty.protobuf.LazyStringArrayList.emptyList();
private void ensureNameServiceIdsIsMutable() {
if (!nameServiceIds_.isModifiable()) {
nameServiceIds_ = new org.apache.hadoop.thirdparty.protobuf.LazyStringArrayList(nameServiceIds_);
}
bitField0_ |= 0x00000001;
}
/**
* <code>repeated string nameServiceIds = 1;</code>
* @return A list containing the nameServiceIds.
*/
public org.apache.hadoop.thirdparty.protobuf.ProtocolStringList
getNameServiceIdsList() {
nameServiceIds_.makeImmutable();
return nameServiceIds_;
}
/**
* <code>repeated string nameServiceIds = 1;</code>
* @return The count of nameServiceIds.
*/
public int getNameServiceIdsCount() {
return nameServiceIds_.size();
}
/**
* <code>repeated string nameServiceIds = 1;</code>
* @param index The index of the element to return.
* @return The nameServiceIds at the given index.
*/
public java.lang.String getNameServiceIds(int index) {
return nameServiceIds_.get(index);
}
/**
* <code>repeated string nameServiceIds = 1;</code>
* @param index The index of the value to return.
* @return The bytes of the nameServiceIds at the given index.
*/
public org.apache.hadoop.thirdparty.protobuf.ByteString
getNameServiceIdsBytes(int index) {
return nameServiceIds_.getByteString(index);
}
/**
* <code>repeated string nameServiceIds = 1;</code>
* @param index The index to set the value at.
* @param value The nameServiceIds to set.
* @return This builder for chaining.
*/
public Builder setNameServiceIds(
int index, java.lang.String value) {
if (value == null) { throw new NullPointerException(); }
ensureNameServiceIdsIsMutable();
nameServiceIds_.set(index, value);
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>repeated string nameServiceIds = 1;</code>
* @param value The nameServiceIds to add.
* @return This builder for chaining.
*/
public Builder addNameServiceIds(
java.lang.String value) {
if (value == null) { throw new NullPointerException(); }
ensureNameServiceIdsIsMutable();
nameServiceIds_.add(value);
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>repeated string nameServiceIds = 1;</code>
* @param values The nameServiceIds to add.
* @return This builder for chaining.
*/
public Builder addAllNameServiceIds(
java.lang.Iterable<java.lang.String> values) {
ensureNameServiceIdsIsMutable();
org.apache.hadoop.thirdparty.protobuf.AbstractMessageLite.Builder.addAll(
values, nameServiceIds_);
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* <code>repeated string nameServiceIds = 1;</code>
* @return This builder for chaining.
*/
public Builder clearNameServiceIds() {
nameServiceIds_ =
org.apache.hadoop.thirdparty.protobuf.LazyStringArrayList.emptyList();
bitField0_ = (bitField0_ & ~0x00000001);;
onChanged();
return this;
}
/**
* <code>repeated string nameServiceIds = 1;</code>
* @param value The bytes of the nameServiceIds to add.
* @return This builder for chaining.
*/
public Builder addNameServiceIdsBytes(
org.apache.hadoop.thirdparty.protobuf.ByteString value) {
if (value == null) { throw new NullPointerException(); }
ensureNameServiceIdsIsMutable();
nameServiceIds_.add(value);
bitField0_ |= 0x00000001;
onChanged();
return this;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.GetDisabledNameservicesResponseProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.GetDisabledNameservicesResponseProto)
private static final org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesResponseProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesResponseProto();
}
public static org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesResponseProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser<GetDisabledNameservicesResponseProto>
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser<GetDisabledNameservicesResponseProto>() {
@java.lang.Override
public GetDisabledNameservicesResponseProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser<GetDisabledNameservicesResponseProto> parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser<GetDisabledNameservicesResponseProto> getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.federation.protocol.proto.HdfsServerFederationProtos.GetDisabledNameservicesResponseProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_NamenodeMembershipStatsRecordProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_NamenodeMembershipStatsRecordProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_NamenodeMembershipRecordProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_NamenodeMembershipRecordProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_FederationNamespaceInfoProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_FederationNamespaceInfoProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_GetNamenodeRegistrationsRequestProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_GetNamenodeRegistrationsRequestProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_GetNamenodeRegistrationsResponseProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_GetNamenodeRegistrationsResponseProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_GetExpiredRegistrationsRequestProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_GetExpiredRegistrationsRequestProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_GetNamespaceInfoRequestProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_GetNamespaceInfoRequestProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_GetNamespaceInfoResponseProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_GetNamespaceInfoResponseProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_UpdateNamenodeRegistrationRequestProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_UpdateNamenodeRegistrationRequestProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_UpdateNamenodeRegistrationResponseProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_UpdateNamenodeRegistrationResponseProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_NamenodeHeartbeatRequestProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_NamenodeHeartbeatRequestProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_NamenodeHeartbeatResponseProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_NamenodeHeartbeatResponseProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_RemoteLocationProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_RemoteLocationProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_MountTableRecordProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_MountTableRecordProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_AddMountTableEntryRequestProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_AddMountTableEntryRequestProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_AddMountTableEntryResponseProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_AddMountTableEntryResponseProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_AddMountTableEntriesRequestProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_AddMountTableEntriesRequestProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_AddMountTableEntriesResponseProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_AddMountTableEntriesResponseProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_UpdateMountTableEntryRequestProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_UpdateMountTableEntryRequestProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_UpdateMountTableEntryResponseProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_UpdateMountTableEntryResponseProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_RemoveMountTableEntryRequestProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_RemoveMountTableEntryRequestProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_RemoveMountTableEntryResponseProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_RemoveMountTableEntryResponseProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_GetMountTableEntriesRequestProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_GetMountTableEntriesRequestProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_GetMountTableEntriesResponseProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_GetMountTableEntriesResponseProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_GetDestinationRequestProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_GetDestinationRequestProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_GetDestinationResponseProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_GetDestinationResponseProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_StateStoreVersionRecordProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_StateStoreVersionRecordProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_RouterRecordProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_RouterRecordProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_GetRouterRegistrationRequestProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_GetRouterRegistrationRequestProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_GetRouterRegistrationResponseProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_GetRouterRegistrationResponseProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_GetRouterRegistrationsRequestProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_GetRouterRegistrationsRequestProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_GetRouterRegistrationsResponseProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_GetRouterRegistrationsResponseProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_RouterHeartbeatRequestProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_RouterHeartbeatRequestProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_RouterHeartbeatResponseProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_RouterHeartbeatResponseProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_RefreshMountTableEntriesRequestProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_RefreshMountTableEntriesRequestProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_RefreshMountTableEntriesResponseProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_RefreshMountTableEntriesResponseProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_RefreshSuperUserGroupsConfigurationRequestProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_RefreshSuperUserGroupsConfigurationRequestProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_RefreshSuperUserGroupsConfigurationResponseProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_RefreshSuperUserGroupsConfigurationResponseProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_EnterSafeModeRequestProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_EnterSafeModeRequestProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_EnterSafeModeResponseProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_EnterSafeModeResponseProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_LeaveSafeModeRequestProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_LeaveSafeModeRequestProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_LeaveSafeModeResponseProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_LeaveSafeModeResponseProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_GetSafeModeRequestProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_GetSafeModeRequestProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_GetSafeModeResponseProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_GetSafeModeResponseProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_DisabledNameserviceRecordProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_DisabledNameserviceRecordProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_DisableNameserviceRequestProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_DisableNameserviceRequestProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_DisableNameserviceResponseProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_DisableNameserviceResponseProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_EnableNameserviceRequestProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_EnableNameserviceRequestProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_EnableNameserviceResponseProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_EnableNameserviceResponseProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_GetDisabledNameservicesRequestProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_GetDisabledNameservicesRequestProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_GetDisabledNameservicesResponseProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_GetDisabledNameservicesResponseProto_fieldAccessorTable;
public static org.apache.hadoop.thirdparty.protobuf.Descriptors.FileDescriptor
getDescriptor() {
return descriptor;
}
private static org.apache.hadoop.thirdparty.protobuf.Descriptors.FileDescriptor
descriptor;
static {
java.lang.String[] descriptorData = {
"\n\030FederationProtocol.proto\022\013hadoop.hdfs\032" +
"\nhdfs.proto\"\355\006\n\"NamenodeMembershipStatsR" +
"ecordProto\022\022\n\ntotalSpace\030\001 \001(\004\022\026\n\016availa" +
"bleSpace\030\002 \001(\004\022\025\n\rprovidedSpace\030\003 \001(\004\022\022\n" +
"\nnumOfFiles\030\n \001(\004\022\023\n\013numOfBlocks\030\013 \001(\004\022\032" +
"\n\022numOfBlocksMissing\030\014 \001(\004\022%\n\035numOfBlock" +
"sPendingReplication\030\r \001(\004\022\"\n\032numOfBlocks" +
"UnderReplicated\030\016 \001(\004\022\"\n\032numOfBlocksPend" +
"ingDeletion\030\017 \001(\004\022\034\n\024numOfActiveDatanode" +
"s\030\024 \001(\r\022\032\n\022numOfDeadDatanodes\030\025 \001(\r\022%\n\035n" +
"umOfDecommissioningDatanodes\030\026 \001(\r\022!\n\031nu" +
"mOfDecomActiveDatanodes\030\027 \001(\r\022\037\n\027numOfDe" +
"comDeadDatanodes\030\030 \001(\r\022\033\n\023numOfStaleData" +
"nodes\030\031 \001(\r\022\'\n\037numOfInMaintenanceLiveDat" +
"aNodes\030\032 \001(\r\022\'\n\037numOfInMaintenanceDeadDa" +
"taNodes\030\033 \001(\r\022)\n!numOfEnteringMaintenanc" +
"eDataNodes\030\034 \001(\r\022\031\n\021corruptFilesCount\030\035 " +
"\001(\r\022\"\n\032scheduledReplicationBlocks\030\036 \001(\004\022" +
"5\n-numberOfMissingBlocksWithReplicationF" +
"actorOne\030\037 \001(\004\0224\n,highestPriorityLowRedu" +
"ndancyReplicatedBlocks\030 \001(\004\022,\n$HighestP" +
"riorityLowRedundancyECBlocks\030! \001(\004\022\027\n\017pe" +
"ndingSPSPaths\030\" \001(\r\022\036\n\026badlyDistributedB" +
"locks\030# \001(\004\"\223\003\n\035NamenodeMembershipRecord" +
"Proto\022\023\n\013dateCreated\030\001 \001(\004\022\024\n\014dateModifi" +
"ed\030\002 \001(\004\022\023\n\013lastContact\030\003 \001(\004\022\020\n\010routerI" +
"d\030\004 \001(\t\022\025\n\rnameserviceId\030\005 \001(\t\022\022\n\nnameno" +
"deId\030\006 \001(\t\022\021\n\tclusterId\030\007 \001(\t\022\023\n\013blockPo" +
"olId\030\010 \001(\t\022\022\n\nwebAddress\030\t \001(\t\022\022\n\nrpcAdd" +
"ress\030\n \001(\t\022\026\n\016serviceAddress\030\013 \001(\t\022\027\n\017li" +
"felineAddress\030\014 \001(\t\022\r\n\005state\030\r \001(\t\022\022\n\nis" +
"SafeMode\030\016 \001(\010\022>\n\005stats\030\017 \001(\0132/.hadoop.h" +
"dfs.NamenodeMembershipStatsRecordProto\022\021" +
"\n\twebScheme\030\020 \001(\t\"]\n\034FederationNamespace" +
"InfoProto\022\023\n\013blockPoolId\030\001 \001(\t\022\021\n\tcluste" +
"rId\030\002 \001(\t\022\025\n\rnameserviceId\030\003 \001(\t\"f\n$GetN" +
"amenodeRegistrationsRequestProto\022>\n\nmemb" +
"ership\030\001 \001(\0132*.hadoop.hdfs.NamenodeMembe" +
"rshipRecordProto\"p\n%GetNamenodeRegistrat" +
"ionsResponseProto\022G\n\023namenodeMemberships" +
"\030\001 \003(\0132*.hadoop.hdfs.NamenodeMembershipR" +
"ecordProto\"%\n#GetExpiredRegistrationsReq" +
"uestProto\"\036\n\034GetNamespaceInfoRequestProt" +
"o\"b\n\035GetNamespaceInfoResponseProto\022A\n\016na" +
"mespaceInfos\030\001 \003(\0132).hadoop.hdfs.Federat" +
"ionNamespaceInfoProto\"b\n&UpdateNamenodeR" +
"egistrationRequestProto\022\025\n\rnameserviceId" +
"\030\001 \001(\t\022\022\n\nnamenodeId\030\002 \001(\t\022\r\n\005state\030\003 \001(" +
"\t\"9\n\'UpdateNamenodeRegistrationResponseP" +
"roto\022\016\n\006status\030\001 \001(\010\"g\n\035NamenodeHeartbea" +
"tRequestProto\022F\n\022namenodeMembership\030\001 \001(" +
"\0132*.hadoop.hdfs.NamenodeMembershipRecord" +
"Proto\"0\n\036NamenodeHeartbeatResponseProto\022" +
"\016\n\006status\030\001 \001(\010\":\n\023RemoteLocationProto\022\025" +
"\n\rnameserviceId\030\001 \001(\t\022\014\n\004path\030\002 \001(\t\"\306\003\n\025" +
"MountTableRecordProto\022\017\n\007srcPath\030\001 \001(\t\0226" +
"\n\014destinations\030\002 \003(\0132 .hadoop.hdfs.Remot" +
"eLocationProto\022\023\n\013dateCreated\030\003 \001(\004\022\024\n\014d" +
"ateModified\030\004 \001(\004\022\027\n\010readOnly\030\005 \001(\010:\005fal" +
"se\022E\n\tdestOrder\030\006 \001(\0162,.hadoop.hdfs.Moun" +
"tTableRecordProto.DestOrder:\004HASH\022\021\n\town" +
"erName\030\n \001(\t\022\021\n\tgroupName\030\013 \001(\t\022\014\n\004mode\030" +
"\014 \001(\005\022+\n\005quota\030\r \001(\0132\034.hadoop.hdfs.Quota" +
"UsageProto\022\034\n\rfaultTolerant\030\016 \001(\010:\005false" +
"\"Z\n\tDestOrder\022\010\n\004HASH\020\000\022\t\n\005LOCAL\020\001\022\n\n\006RA" +
"NDOM\020\002\022\014\n\010HASH_ALL\020\003\022\t\n\005SPACE\020\004\022\023\n\017LEADE" +
"R_FOLLOWER\020\005\"S\n\036AddMountTableEntryReques" +
"tProto\0221\n\005entry\030\001 \001(\0132\".hadoop.hdfs.Moun" +
"tTableRecordProto\"1\n\037AddMountTableEntryR" +
"esponseProto\022\016\n\006status\030\001 \001(\010\"U\n AddMount" +
"TableEntriesRequestProto\0221\n\005entry\030\001 \003(\0132" +
"\".hadoop.hdfs.MountTableRecordProto\"N\n!A" +
"ddMountTableEntriesResponseProto\022\016\n\006stat" +
"us\030\001 \001(\010\022\031\n\021failedEntriesKeys\030\002 \003(\t\"V\n!U" +
"pdateMountTableEntryRequestProto\0221\n\005entr" +
"y\030\001 \001(\0132\".hadoop.hdfs.MountTableRecordPr" +
"oto\"4\n\"UpdateMountTableEntryResponseProt" +
"o\022\016\n\006status\030\001 \001(\010\"4\n!RemoveMountTableEnt" +
"ryRequestProto\022\017\n\007srcPath\030\001 \001(\t\"4\n\"Remov" +
"eMountTableEntryResponseProto\022\016\n\006status\030" +
"\001 \001(\010\"3\n GetMountTableEntriesRequestProt" +
"o\022\017\n\007srcPath\030\001 \001(\t\"k\n!GetMountTableEntri" +
"esResponseProto\0223\n\007entries\030\001 \003(\0132\".hadoo" +
"p.hdfs.MountTableRecordProto\022\021\n\ttimestam" +
"p\030\002 \001(\004\"-\n\032GetDestinationRequestProto\022\017\n" +
"\007srcPath\030\001 \001(\t\"3\n\033GetDestinationResponse" +
"Proto\022\024\n\014destinations\030\001 \003(\t\"T\n\034StateStor" +
"eVersionRecordProto\022\031\n\021membershipVersion" +
"\030\001 \001(\004\022\031\n\021mountTableVersion\030\002 \001(\004\"\366\001\n\021Ro" +
"uterRecordProto\022\023\n\013dateCreated\030\001 \001(\004\022\024\n\014" +
"dateModified\030\002 \001(\004\022\017\n\007address\030\003 \001(\t\022\016\n\006s" +
"tatus\030\004 \001(\t\022D\n\021stateStoreVersion\030\005 \001(\0132)" +
".hadoop.hdfs.StateStoreVersionRecordProt" +
"o\022\017\n\007version\030\006 \001(\t\022\023\n\013compileInfo\030\007 \001(\t\022" +
"\023\n\013dateStarted\030\010 \001(\004\022\024\n\014adminAddress\030\t \001" +
"(\t\"5\n!GetRouterRegistrationRequestProto\022" +
"\020\n\010routerId\030\001 \001(\t\"T\n\"GetRouterRegistrati" +
"onResponseProto\022.\n\006router\030\001 \001(\0132\036.hadoop" +
".hdfs.RouterRecordProto\"$\n\"GetRouterRegi" +
"strationsRequestProto\"i\n#GetRouterRegist" +
"rationsResponseProto\022\021\n\ttimestamp\030\001 \001(\004\022" +
"/\n\007routers\030\002 \003(\0132\036.hadoop.hdfs.RouterRec" +
"ordProto\"M\n\033RouterHeartbeatRequestProto\022" +
".\n\006router\030\001 \001(\0132\036.hadoop.hdfs.RouterReco" +
"rdProto\".\n\034RouterHeartbeatResponseProto\022" +
"\016\n\006status\030\001 \001(\010\"&\n$RefreshMountTableEntr" +
"iesRequestProto\"7\n%RefreshMountTableEntr" +
"iesResponseProto\022\016\n\006result\030\001 \001(\010\"1\n/Refr" +
"eshSuperUserGroupsConfigurationRequestPr" +
"oto\"B\n0RefreshSuperUserGroupsConfigurati" +
"onResponseProto\022\016\n\006status\030\001 \001(\010\"\033\n\031Enter" +
"SafeModeRequestProto\",\n\032EnterSafeModeRes" +
"ponseProto\022\016\n\006status\030\001 \001(\010\"\033\n\031LeaveSafeM" +
"odeRequestProto\",\n\032LeaveSafeModeResponse" +
"Proto\022\016\n\006status\030\001 \001(\010\"\031\n\027GetSafeModeRequ" +
"estProto\"0\n\030GetSafeModeResponseProto\022\024\n\014" +
"isInSafeMode\030\001 \001(\010\"b\n\036DisabledNameservic" +
"eRecordProto\022\023\n\013dateCreated\030\001 \001(\004\022\024\n\014dat" +
"eModified\030\002 \001(\004\022\025\n\rnameServiceId\030\003 \001(\t\"7" +
"\n\036DisableNameserviceRequestProto\022\025\n\rname" +
"ServiceId\030\001 \001(\t\"1\n\037DisableNameserviceRes" +
"ponseProto\022\016\n\006status\030\001 \001(\010\"6\n\035EnableName" +
"serviceRequestProto\022\025\n\rnameServiceId\030\001 \001" +
"(\t\"0\n\036EnableNameserviceResponseProto\022\016\n\006" +
"status\030\001 \001(\010\"%\n#GetDisabledNameservicesR" +
"equestProto\">\n$GetDisabledNameservicesRe" +
"sponseProto\022\026\n\016nameServiceIds\030\001 \003(\tBT\n0o" +
"rg.apache.hadoop.hdfs.federation.protoco" +
"l.protoB\032HdfsServerFederationProtos\210\001\001\240\001" +
"\001"
};
descriptor = org.apache.hadoop.thirdparty.protobuf.Descriptors.FileDescriptor
.internalBuildGeneratedFileFrom(descriptorData,
new org.apache.hadoop.thirdparty.protobuf.Descriptors.FileDescriptor[] {
org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.getDescriptor(),
});
internal_static_hadoop_hdfs_NamenodeMembershipStatsRecordProto_descriptor =
getDescriptor().getMessageTypes().get(0);
internal_static_hadoop_hdfs_NamenodeMembershipStatsRecordProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_NamenodeMembershipStatsRecordProto_descriptor,
new java.lang.String[] { "TotalSpace", "AvailableSpace", "ProvidedSpace", "NumOfFiles", "NumOfBlocks", "NumOfBlocksMissing", "NumOfBlocksPendingReplication", "NumOfBlocksUnderReplicated", "NumOfBlocksPendingDeletion", "NumOfActiveDatanodes", "NumOfDeadDatanodes", "NumOfDecommissioningDatanodes", "NumOfDecomActiveDatanodes", "NumOfDecomDeadDatanodes", "NumOfStaleDatanodes", "NumOfInMaintenanceLiveDataNodes", "NumOfInMaintenanceDeadDataNodes", "NumOfEnteringMaintenanceDataNodes", "CorruptFilesCount", "ScheduledReplicationBlocks", "NumberOfMissingBlocksWithReplicationFactorOne", "HighestPriorityLowRedundancyReplicatedBlocks", "HighestPriorityLowRedundancyECBlocks", "PendingSPSPaths", "BadlyDistributedBlocks", });
internal_static_hadoop_hdfs_NamenodeMembershipRecordProto_descriptor =
getDescriptor().getMessageTypes().get(1);
internal_static_hadoop_hdfs_NamenodeMembershipRecordProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_NamenodeMembershipRecordProto_descriptor,
new java.lang.String[] { "DateCreated", "DateModified", "LastContact", "RouterId", "NameserviceId", "NamenodeId", "ClusterId", "BlockPoolId", "WebAddress", "RpcAddress", "ServiceAddress", "LifelineAddress", "State", "IsSafeMode", "Stats", "WebScheme", });
internal_static_hadoop_hdfs_FederationNamespaceInfoProto_descriptor =
getDescriptor().getMessageTypes().get(2);
internal_static_hadoop_hdfs_FederationNamespaceInfoProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_FederationNamespaceInfoProto_descriptor,
new java.lang.String[] { "BlockPoolId", "ClusterId", "NameserviceId", });
internal_static_hadoop_hdfs_GetNamenodeRegistrationsRequestProto_descriptor =
getDescriptor().getMessageTypes().get(3);
internal_static_hadoop_hdfs_GetNamenodeRegistrationsRequestProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_GetNamenodeRegistrationsRequestProto_descriptor,
new java.lang.String[] { "Membership", });
internal_static_hadoop_hdfs_GetNamenodeRegistrationsResponseProto_descriptor =
getDescriptor().getMessageTypes().get(4);
internal_static_hadoop_hdfs_GetNamenodeRegistrationsResponseProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_GetNamenodeRegistrationsResponseProto_descriptor,
new java.lang.String[] { "NamenodeMemberships", });
internal_static_hadoop_hdfs_GetExpiredRegistrationsRequestProto_descriptor =
getDescriptor().getMessageTypes().get(5);
internal_static_hadoop_hdfs_GetExpiredRegistrationsRequestProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_GetExpiredRegistrationsRequestProto_descriptor,
new java.lang.String[] { });
internal_static_hadoop_hdfs_GetNamespaceInfoRequestProto_descriptor =
getDescriptor().getMessageTypes().get(6);
internal_static_hadoop_hdfs_GetNamespaceInfoRequestProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_GetNamespaceInfoRequestProto_descriptor,
new java.lang.String[] { });
internal_static_hadoop_hdfs_GetNamespaceInfoResponseProto_descriptor =
getDescriptor().getMessageTypes().get(7);
internal_static_hadoop_hdfs_GetNamespaceInfoResponseProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_GetNamespaceInfoResponseProto_descriptor,
new java.lang.String[] { "NamespaceInfos", });
internal_static_hadoop_hdfs_UpdateNamenodeRegistrationRequestProto_descriptor =
getDescriptor().getMessageTypes().get(8);
internal_static_hadoop_hdfs_UpdateNamenodeRegistrationRequestProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_UpdateNamenodeRegistrationRequestProto_descriptor,
new java.lang.String[] { "NameserviceId", "NamenodeId", "State", });
internal_static_hadoop_hdfs_UpdateNamenodeRegistrationResponseProto_descriptor =
getDescriptor().getMessageTypes().get(9);
internal_static_hadoop_hdfs_UpdateNamenodeRegistrationResponseProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_UpdateNamenodeRegistrationResponseProto_descriptor,
new java.lang.String[] { "Status", });
internal_static_hadoop_hdfs_NamenodeHeartbeatRequestProto_descriptor =
getDescriptor().getMessageTypes().get(10);
internal_static_hadoop_hdfs_NamenodeHeartbeatRequestProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_NamenodeHeartbeatRequestProto_descriptor,
new java.lang.String[] { "NamenodeMembership", });
internal_static_hadoop_hdfs_NamenodeHeartbeatResponseProto_descriptor =
getDescriptor().getMessageTypes().get(11);
internal_static_hadoop_hdfs_NamenodeHeartbeatResponseProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_NamenodeHeartbeatResponseProto_descriptor,
new java.lang.String[] { "Status", });
internal_static_hadoop_hdfs_RemoteLocationProto_descriptor =
getDescriptor().getMessageTypes().get(12);
internal_static_hadoop_hdfs_RemoteLocationProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_RemoteLocationProto_descriptor,
new java.lang.String[] { "NameserviceId", "Path", });
internal_static_hadoop_hdfs_MountTableRecordProto_descriptor =
getDescriptor().getMessageTypes().get(13);
internal_static_hadoop_hdfs_MountTableRecordProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_MountTableRecordProto_descriptor,
new java.lang.String[] { "SrcPath", "Destinations", "DateCreated", "DateModified", "ReadOnly", "DestOrder", "OwnerName", "GroupName", "Mode", "Quota", "FaultTolerant", });
internal_static_hadoop_hdfs_AddMountTableEntryRequestProto_descriptor =
getDescriptor().getMessageTypes().get(14);
internal_static_hadoop_hdfs_AddMountTableEntryRequestProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_AddMountTableEntryRequestProto_descriptor,
new java.lang.String[] { "Entry", });
internal_static_hadoop_hdfs_AddMountTableEntryResponseProto_descriptor =
getDescriptor().getMessageTypes().get(15);
internal_static_hadoop_hdfs_AddMountTableEntryResponseProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_AddMountTableEntryResponseProto_descriptor,
new java.lang.String[] { "Status", });
internal_static_hadoop_hdfs_AddMountTableEntriesRequestProto_descriptor =
getDescriptor().getMessageTypes().get(16);
internal_static_hadoop_hdfs_AddMountTableEntriesRequestProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_AddMountTableEntriesRequestProto_descriptor,
new java.lang.String[] { "Entry", });
internal_static_hadoop_hdfs_AddMountTableEntriesResponseProto_descriptor =
getDescriptor().getMessageTypes().get(17);
internal_static_hadoop_hdfs_AddMountTableEntriesResponseProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_AddMountTableEntriesResponseProto_descriptor,
new java.lang.String[] { "Status", "FailedEntriesKeys", });
internal_static_hadoop_hdfs_UpdateMountTableEntryRequestProto_descriptor =
getDescriptor().getMessageTypes().get(18);
internal_static_hadoop_hdfs_UpdateMountTableEntryRequestProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_UpdateMountTableEntryRequestProto_descriptor,
new java.lang.String[] { "Entry", });
internal_static_hadoop_hdfs_UpdateMountTableEntryResponseProto_descriptor =
getDescriptor().getMessageTypes().get(19);
internal_static_hadoop_hdfs_UpdateMountTableEntryResponseProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_UpdateMountTableEntryResponseProto_descriptor,
new java.lang.String[] { "Status", });
internal_static_hadoop_hdfs_RemoveMountTableEntryRequestProto_descriptor =
getDescriptor().getMessageTypes().get(20);
internal_static_hadoop_hdfs_RemoveMountTableEntryRequestProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_RemoveMountTableEntryRequestProto_descriptor,
new java.lang.String[] { "SrcPath", });
internal_static_hadoop_hdfs_RemoveMountTableEntryResponseProto_descriptor =
getDescriptor().getMessageTypes().get(21);
internal_static_hadoop_hdfs_RemoveMountTableEntryResponseProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_RemoveMountTableEntryResponseProto_descriptor,
new java.lang.String[] { "Status", });
internal_static_hadoop_hdfs_GetMountTableEntriesRequestProto_descriptor =
getDescriptor().getMessageTypes().get(22);
internal_static_hadoop_hdfs_GetMountTableEntriesRequestProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_GetMountTableEntriesRequestProto_descriptor,
new java.lang.String[] { "SrcPath", });
internal_static_hadoop_hdfs_GetMountTableEntriesResponseProto_descriptor =
getDescriptor().getMessageTypes().get(23);
internal_static_hadoop_hdfs_GetMountTableEntriesResponseProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_GetMountTableEntriesResponseProto_descriptor,
new java.lang.String[] { "Entries", "Timestamp", });
internal_static_hadoop_hdfs_GetDestinationRequestProto_descriptor =
getDescriptor().getMessageTypes().get(24);
internal_static_hadoop_hdfs_GetDestinationRequestProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_GetDestinationRequestProto_descriptor,
new java.lang.String[] { "SrcPath", });
internal_static_hadoop_hdfs_GetDestinationResponseProto_descriptor =
getDescriptor().getMessageTypes().get(25);
internal_static_hadoop_hdfs_GetDestinationResponseProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_GetDestinationResponseProto_descriptor,
new java.lang.String[] { "Destinations", });
internal_static_hadoop_hdfs_StateStoreVersionRecordProto_descriptor =
getDescriptor().getMessageTypes().get(26);
internal_static_hadoop_hdfs_StateStoreVersionRecordProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_StateStoreVersionRecordProto_descriptor,
new java.lang.String[] { "MembershipVersion", "MountTableVersion", });
internal_static_hadoop_hdfs_RouterRecordProto_descriptor =
getDescriptor().getMessageTypes().get(27);
internal_static_hadoop_hdfs_RouterRecordProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_RouterRecordProto_descriptor,
new java.lang.String[] { "DateCreated", "DateModified", "Address", "Status", "StateStoreVersion", "Version", "CompileInfo", "DateStarted", "AdminAddress", });
internal_static_hadoop_hdfs_GetRouterRegistrationRequestProto_descriptor =
getDescriptor().getMessageTypes().get(28);
internal_static_hadoop_hdfs_GetRouterRegistrationRequestProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_GetRouterRegistrationRequestProto_descriptor,
new java.lang.String[] { "RouterId", });
internal_static_hadoop_hdfs_GetRouterRegistrationResponseProto_descriptor =
getDescriptor().getMessageTypes().get(29);
internal_static_hadoop_hdfs_GetRouterRegistrationResponseProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_GetRouterRegistrationResponseProto_descriptor,
new java.lang.String[] { "Router", });
internal_static_hadoop_hdfs_GetRouterRegistrationsRequestProto_descriptor =
getDescriptor().getMessageTypes().get(30);
internal_static_hadoop_hdfs_GetRouterRegistrationsRequestProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_GetRouterRegistrationsRequestProto_descriptor,
new java.lang.String[] { });
internal_static_hadoop_hdfs_GetRouterRegistrationsResponseProto_descriptor =
getDescriptor().getMessageTypes().get(31);
internal_static_hadoop_hdfs_GetRouterRegistrationsResponseProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_GetRouterRegistrationsResponseProto_descriptor,
new java.lang.String[] { "Timestamp", "Routers", });
internal_static_hadoop_hdfs_RouterHeartbeatRequestProto_descriptor =
getDescriptor().getMessageTypes().get(32);
internal_static_hadoop_hdfs_RouterHeartbeatRequestProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_RouterHeartbeatRequestProto_descriptor,
new java.lang.String[] { "Router", });
internal_static_hadoop_hdfs_RouterHeartbeatResponseProto_descriptor =
getDescriptor().getMessageTypes().get(33);
internal_static_hadoop_hdfs_RouterHeartbeatResponseProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_RouterHeartbeatResponseProto_descriptor,
new java.lang.String[] { "Status", });
internal_static_hadoop_hdfs_RefreshMountTableEntriesRequestProto_descriptor =
getDescriptor().getMessageTypes().get(34);
internal_static_hadoop_hdfs_RefreshMountTableEntriesRequestProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_RefreshMountTableEntriesRequestProto_descriptor,
new java.lang.String[] { });
internal_static_hadoop_hdfs_RefreshMountTableEntriesResponseProto_descriptor =
getDescriptor().getMessageTypes().get(35);
internal_static_hadoop_hdfs_RefreshMountTableEntriesResponseProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_RefreshMountTableEntriesResponseProto_descriptor,
new java.lang.String[] { "Result", });
internal_static_hadoop_hdfs_RefreshSuperUserGroupsConfigurationRequestProto_descriptor =
getDescriptor().getMessageTypes().get(36);
internal_static_hadoop_hdfs_RefreshSuperUserGroupsConfigurationRequestProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_RefreshSuperUserGroupsConfigurationRequestProto_descriptor,
new java.lang.String[] { });
internal_static_hadoop_hdfs_RefreshSuperUserGroupsConfigurationResponseProto_descriptor =
getDescriptor().getMessageTypes().get(37);
internal_static_hadoop_hdfs_RefreshSuperUserGroupsConfigurationResponseProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_RefreshSuperUserGroupsConfigurationResponseProto_descriptor,
new java.lang.String[] { "Status", });
internal_static_hadoop_hdfs_EnterSafeModeRequestProto_descriptor =
getDescriptor().getMessageTypes().get(38);
internal_static_hadoop_hdfs_EnterSafeModeRequestProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_EnterSafeModeRequestProto_descriptor,
new java.lang.String[] { });
internal_static_hadoop_hdfs_EnterSafeModeResponseProto_descriptor =
getDescriptor().getMessageTypes().get(39);
internal_static_hadoop_hdfs_EnterSafeModeResponseProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_EnterSafeModeResponseProto_descriptor,
new java.lang.String[] { "Status", });
internal_static_hadoop_hdfs_LeaveSafeModeRequestProto_descriptor =
getDescriptor().getMessageTypes().get(40);
internal_static_hadoop_hdfs_LeaveSafeModeRequestProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_LeaveSafeModeRequestProto_descriptor,
new java.lang.String[] { });
internal_static_hadoop_hdfs_LeaveSafeModeResponseProto_descriptor =
getDescriptor().getMessageTypes().get(41);
internal_static_hadoop_hdfs_LeaveSafeModeResponseProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_LeaveSafeModeResponseProto_descriptor,
new java.lang.String[] { "Status", });
internal_static_hadoop_hdfs_GetSafeModeRequestProto_descriptor =
getDescriptor().getMessageTypes().get(42);
internal_static_hadoop_hdfs_GetSafeModeRequestProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_GetSafeModeRequestProto_descriptor,
new java.lang.String[] { });
internal_static_hadoop_hdfs_GetSafeModeResponseProto_descriptor =
getDescriptor().getMessageTypes().get(43);
internal_static_hadoop_hdfs_GetSafeModeResponseProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_GetSafeModeResponseProto_descriptor,
new java.lang.String[] { "IsInSafeMode", });
internal_static_hadoop_hdfs_DisabledNameserviceRecordProto_descriptor =
getDescriptor().getMessageTypes().get(44);
internal_static_hadoop_hdfs_DisabledNameserviceRecordProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_DisabledNameserviceRecordProto_descriptor,
new java.lang.String[] { "DateCreated", "DateModified", "NameServiceId", });
internal_static_hadoop_hdfs_DisableNameserviceRequestProto_descriptor =
getDescriptor().getMessageTypes().get(45);
internal_static_hadoop_hdfs_DisableNameserviceRequestProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_DisableNameserviceRequestProto_descriptor,
new java.lang.String[] { "NameServiceId", });
internal_static_hadoop_hdfs_DisableNameserviceResponseProto_descriptor =
getDescriptor().getMessageTypes().get(46);
internal_static_hadoop_hdfs_DisableNameserviceResponseProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_DisableNameserviceResponseProto_descriptor,
new java.lang.String[] { "Status", });
internal_static_hadoop_hdfs_EnableNameserviceRequestProto_descriptor =
getDescriptor().getMessageTypes().get(47);
internal_static_hadoop_hdfs_EnableNameserviceRequestProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_EnableNameserviceRequestProto_descriptor,
new java.lang.String[] { "NameServiceId", });
internal_static_hadoop_hdfs_EnableNameserviceResponseProto_descriptor =
getDescriptor().getMessageTypes().get(48);
internal_static_hadoop_hdfs_EnableNameserviceResponseProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_EnableNameserviceResponseProto_descriptor,
new java.lang.String[] { "Status", });
internal_static_hadoop_hdfs_GetDisabledNameservicesRequestProto_descriptor =
getDescriptor().getMessageTypes().get(49);
internal_static_hadoop_hdfs_GetDisabledNameservicesRequestProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_GetDisabledNameservicesRequestProto_descriptor,
new java.lang.String[] { });
internal_static_hadoop_hdfs_GetDisabledNameservicesResponseProto_descriptor =
getDescriptor().getMessageTypes().get(50);
internal_static_hadoop_hdfs_GetDisabledNameservicesResponseProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_GetDisabledNameservicesResponseProto_descriptor,
new java.lang.String[] { "NameServiceIds", });
org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.getDescriptor();
}
// @@protoc_insertion_point(outer_class_scope)
}